var/home/core/zuul-output/0000755000175000017500000000000015072146151014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015072161223015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005160455015072161214017702 0ustar rootrootOct 10 09:10:55 crc systemd[1]: Starting Kubernetes Kubelet... Oct 10 09:10:55 crc restorecon[4631]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:55 crc restorecon[4631]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 09:10:56 crc restorecon[4631]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 10 09:10:57 crc kubenswrapper[4669]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 09:10:57 crc kubenswrapper[4669]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 10 09:10:57 crc kubenswrapper[4669]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 09:10:57 crc kubenswrapper[4669]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 09:10:57 crc kubenswrapper[4669]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 10 09:10:57 crc kubenswrapper[4669]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.550989 4669 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556938 4669 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556960 4669 feature_gate.go:330] unrecognized feature gate: Example Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556966 4669 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556971 4669 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556976 4669 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556981 4669 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556986 4669 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556991 4669 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556995 4669 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.556999 4669 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557004 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557008 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557013 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557019 4669 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557024 4669 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557029 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557033 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557046 4669 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557051 4669 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557055 4669 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557059 4669 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557064 4669 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557068 4669 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557072 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557077 4669 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557081 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557085 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557091 4669 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557096 4669 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557101 4669 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557106 4669 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557110 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557115 4669 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557119 4669 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557123 4669 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557128 4669 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557135 4669 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557214 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557219 4669 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557226 4669 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557233 4669 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557238 4669 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557243 4669 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557248 4669 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557252 4669 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557258 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557262 4669 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557267 4669 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557271 4669 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557276 4669 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557281 4669 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557286 4669 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557290 4669 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557295 4669 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557300 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557304 4669 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557308 4669 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557313 4669 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557317 4669 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557322 4669 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557326 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557330 4669 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557335 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557339 4669 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557344 4669 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557349 4669 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557364 4669 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557374 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557378 4669 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557383 4669 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.557387 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557481 4669 flags.go:64] FLAG: --address="0.0.0.0" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557493 4669 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557506 4669 flags.go:64] FLAG: --anonymous-auth="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557512 4669 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557530 4669 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557541 4669 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557549 4669 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557556 4669 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557562 4669 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557567 4669 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557573 4669 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557578 4669 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557583 4669 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557607 4669 flags.go:64] FLAG: --cgroup-root="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557612 4669 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557618 4669 flags.go:64] FLAG: --client-ca-file="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557625 4669 flags.go:64] FLAG: --cloud-config="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557631 4669 flags.go:64] FLAG: --cloud-provider="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557636 4669 flags.go:64] FLAG: --cluster-dns="[]" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557659 4669 flags.go:64] FLAG: --cluster-domain="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557664 4669 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557668 4669 flags.go:64] FLAG: --config-dir="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557672 4669 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557677 4669 flags.go:64] FLAG: --container-log-max-files="5" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557683 4669 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557687 4669 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557691 4669 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557695 4669 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557700 4669 flags.go:64] FLAG: --contention-profiling="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557704 4669 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557708 4669 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557712 4669 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557716 4669 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557722 4669 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557727 4669 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557731 4669 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557735 4669 flags.go:64] FLAG: --enable-load-reader="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557739 4669 flags.go:64] FLAG: --enable-server="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557743 4669 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557753 4669 flags.go:64] FLAG: --event-burst="100" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557758 4669 flags.go:64] FLAG: --event-qps="50" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557762 4669 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557766 4669 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557770 4669 flags.go:64] FLAG: --eviction-hard="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557776 4669 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557781 4669 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557785 4669 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557789 4669 flags.go:64] FLAG: --eviction-soft="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557793 4669 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557798 4669 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557802 4669 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557807 4669 flags.go:64] FLAG: --experimental-mounter-path="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557811 4669 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557815 4669 flags.go:64] FLAG: --fail-swap-on="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557819 4669 flags.go:64] FLAG: --feature-gates="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557824 4669 flags.go:64] FLAG: --file-check-frequency="20s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557828 4669 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557833 4669 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557837 4669 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557842 4669 flags.go:64] FLAG: --healthz-port="10248" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557846 4669 flags.go:64] FLAG: --help="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557851 4669 flags.go:64] FLAG: --hostname-override="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557855 4669 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557859 4669 flags.go:64] FLAG: --http-check-frequency="20s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557863 4669 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557867 4669 flags.go:64] FLAG: --image-credential-provider-config="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557871 4669 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557875 4669 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557879 4669 flags.go:64] FLAG: --image-service-endpoint="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557883 4669 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557887 4669 flags.go:64] FLAG: --kube-api-burst="100" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557891 4669 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557895 4669 flags.go:64] FLAG: --kube-api-qps="50" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557900 4669 flags.go:64] FLAG: --kube-reserved="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557904 4669 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557908 4669 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557913 4669 flags.go:64] FLAG: --kubelet-cgroups="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557917 4669 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557921 4669 flags.go:64] FLAG: --lock-file="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557925 4669 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557929 4669 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557933 4669 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557939 4669 flags.go:64] FLAG: --log-json-split-stream="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557943 4669 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557947 4669 flags.go:64] FLAG: --log-text-split-stream="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557951 4669 flags.go:64] FLAG: --logging-format="text" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557956 4669 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557961 4669 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557966 4669 flags.go:64] FLAG: --manifest-url="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557972 4669 flags.go:64] FLAG: --manifest-url-header="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557979 4669 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557984 4669 flags.go:64] FLAG: --max-open-files="1000000" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557990 4669 flags.go:64] FLAG: --max-pods="110" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.557996 4669 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558000 4669 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558004 4669 flags.go:64] FLAG: --memory-manager-policy="None" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558009 4669 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558013 4669 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558018 4669 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558022 4669 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558032 4669 flags.go:64] FLAG: --node-status-max-images="50" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558037 4669 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558041 4669 flags.go:64] FLAG: --oom-score-adj="-999" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558045 4669 flags.go:64] FLAG: --pod-cidr="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558049 4669 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558056 4669 flags.go:64] FLAG: --pod-manifest-path="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558060 4669 flags.go:64] FLAG: --pod-max-pids="-1" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558064 4669 flags.go:64] FLAG: --pods-per-core="0" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558069 4669 flags.go:64] FLAG: --port="10250" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558073 4669 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558078 4669 flags.go:64] FLAG: --provider-id="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558082 4669 flags.go:64] FLAG: --qos-reserved="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558086 4669 flags.go:64] FLAG: --read-only-port="10255" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558091 4669 flags.go:64] FLAG: --register-node="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558095 4669 flags.go:64] FLAG: --register-schedulable="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558099 4669 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558106 4669 flags.go:64] FLAG: --registry-burst="10" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558110 4669 flags.go:64] FLAG: --registry-qps="5" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558114 4669 flags.go:64] FLAG: --reserved-cpus="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558118 4669 flags.go:64] FLAG: --reserved-memory="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558123 4669 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558127 4669 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558131 4669 flags.go:64] FLAG: --rotate-certificates="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558135 4669 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558139 4669 flags.go:64] FLAG: --runonce="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558143 4669 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558148 4669 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558152 4669 flags.go:64] FLAG: --seccomp-default="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558156 4669 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558160 4669 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558164 4669 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558169 4669 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558174 4669 flags.go:64] FLAG: --storage-driver-password="root" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558178 4669 flags.go:64] FLAG: --storage-driver-secure="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558182 4669 flags.go:64] FLAG: --storage-driver-table="stats" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558186 4669 flags.go:64] FLAG: --storage-driver-user="root" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558190 4669 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558195 4669 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558199 4669 flags.go:64] FLAG: --system-cgroups="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558203 4669 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558209 4669 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558214 4669 flags.go:64] FLAG: --tls-cert-file="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558218 4669 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558232 4669 flags.go:64] FLAG: --tls-min-version="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558236 4669 flags.go:64] FLAG: --tls-private-key-file="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558240 4669 flags.go:64] FLAG: --topology-manager-policy="none" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558244 4669 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558249 4669 flags.go:64] FLAG: --topology-manager-scope="container" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558253 4669 flags.go:64] FLAG: --v="2" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558259 4669 flags.go:64] FLAG: --version="false" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558265 4669 flags.go:64] FLAG: --vmodule="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558270 4669 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558275 4669 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558370 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558374 4669 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558378 4669 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558383 4669 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558388 4669 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558393 4669 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558397 4669 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558400 4669 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558403 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558407 4669 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558411 4669 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558415 4669 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558418 4669 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558422 4669 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558425 4669 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558429 4669 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558432 4669 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558440 4669 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558443 4669 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558448 4669 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558453 4669 feature_gate.go:330] unrecognized feature gate: Example Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558456 4669 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558460 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558463 4669 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558467 4669 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558470 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558474 4669 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558477 4669 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558481 4669 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558484 4669 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558488 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558492 4669 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558496 4669 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558499 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558503 4669 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558506 4669 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558510 4669 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558514 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558518 4669 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558523 4669 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558527 4669 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558531 4669 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558537 4669 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558541 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558544 4669 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558548 4669 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558552 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558555 4669 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558559 4669 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558565 4669 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558568 4669 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558572 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558576 4669 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558583 4669 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558586 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558608 4669 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558612 4669 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558617 4669 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558620 4669 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558624 4669 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558628 4669 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558631 4669 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558635 4669 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558639 4669 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558642 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558646 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558649 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558653 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558657 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558660 4669 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.558664 4669 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.558676 4669 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.570460 4669 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.570516 4669 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570674 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570696 4669 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570705 4669 feature_gate.go:330] unrecognized feature gate: Example Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570715 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570727 4669 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570738 4669 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570747 4669 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570756 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570765 4669 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570774 4669 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570785 4669 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570797 4669 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570805 4669 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570815 4669 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570824 4669 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570832 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570840 4669 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570848 4669 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570856 4669 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570864 4669 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570872 4669 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570880 4669 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570888 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570897 4669 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570907 4669 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570917 4669 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570927 4669 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570937 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570947 4669 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570959 4669 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570967 4669 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570975 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570982 4669 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570990 4669 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.570999 4669 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571007 4669 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571015 4669 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571023 4669 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571033 4669 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571043 4669 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571052 4669 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571060 4669 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571068 4669 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571076 4669 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571083 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571091 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571101 4669 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571111 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571121 4669 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571131 4669 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571141 4669 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571150 4669 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571160 4669 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571170 4669 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571180 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571191 4669 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571201 4669 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571211 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571221 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571230 4669 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571238 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571245 4669 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571253 4669 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571260 4669 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571268 4669 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571275 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571283 4669 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571293 4669 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571302 4669 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571310 4669 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571321 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.571335 4669 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571658 4669 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571676 4669 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571685 4669 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571693 4669 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571703 4669 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571711 4669 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571719 4669 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571727 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571738 4669 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571748 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571758 4669 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571771 4669 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571785 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571796 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571809 4669 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571820 4669 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571831 4669 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571842 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571853 4669 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571862 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571870 4669 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571878 4669 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571886 4669 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571893 4669 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571902 4669 feature_gate.go:330] unrecognized feature gate: Example Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571910 4669 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571917 4669 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571925 4669 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571932 4669 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571940 4669 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571948 4669 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571956 4669 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571964 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571972 4669 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571981 4669 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.571992 4669 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572002 4669 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572012 4669 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572022 4669 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572032 4669 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572041 4669 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572051 4669 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572061 4669 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572074 4669 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572087 4669 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572098 4669 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572108 4669 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572118 4669 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572126 4669 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572134 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572142 4669 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572149 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572160 4669 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572170 4669 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572180 4669 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572188 4669 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572198 4669 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572208 4669 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572216 4669 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572224 4669 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572232 4669 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572241 4669 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572248 4669 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572257 4669 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572264 4669 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572274 4669 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572284 4669 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572294 4669 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572304 4669 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572314 4669 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.572325 4669 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.572340 4669 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.573889 4669 server.go:940] "Client rotation is on, will bootstrap in background" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.578745 4669 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.578858 4669 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.581277 4669 server.go:997] "Starting client certificate rotation" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.581315 4669 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.582137 4669 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-21 18:00:16.687340819 +0000 UTC Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.582232 4669 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1016h49m19.105110614s for next certificate rotation Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.605042 4669 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.606939 4669 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.628887 4669 log.go:25] "Validated CRI v1 runtime API" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.661718 4669 log.go:25] "Validated CRI v1 image API" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.664013 4669 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.671741 4669 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-10-09-04-38-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.671776 4669 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.690016 4669 manager.go:217] Machine: {Timestamp:2025-10-10 09:10:57.685553792 +0000 UTC m=+0.701572574 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:fd4c383f-0ffa-443a-a38f-89337f2bc2e5 BootID:dc51e5f1-4956-4174-82c2-97dec3a7897f Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:96:54:f1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:96:54:f1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:6f:5a:85 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b7:5e:b3 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:76:44:a8 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:84:86:20 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:c3:d9:30 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:26:9f:1d:ba:83:6e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ce:ee:51:b8:5c:99 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.690374 4669 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.690551 4669 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.692063 4669 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.692382 4669 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.692431 4669 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.692780 4669 topology_manager.go:138] "Creating topology manager with none policy" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.692799 4669 container_manager_linux.go:303] "Creating device plugin manager" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.693305 4669 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.693355 4669 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.693670 4669 state_mem.go:36] "Initialized new in-memory state store" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.693791 4669 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.698613 4669 kubelet.go:418] "Attempting to sync node with API server" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.698647 4669 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.698741 4669 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.698773 4669 kubelet.go:324] "Adding apiserver pod source" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.698791 4669 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.704400 4669 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.705747 4669 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.708555 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.708580 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.708660 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.708704 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.709537 4669 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711181 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711205 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711212 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711218 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711233 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711243 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711253 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711268 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711279 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711294 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711320 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.711330 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.714677 4669 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.715827 4669 server.go:1280] "Started kubelet" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.716123 4669 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:57 crc systemd[1]: Started Kubernetes Kubelet. Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.718399 4669 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.718393 4669 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.719665 4669 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.720910 4669 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.721031 4669 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.721083 4669 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 17:42:48.481953621 +0000 UTC Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.721460 4669 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.721663 4669 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1184h31m50.76032118s for next certificate rotation Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.721680 4669 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.721177 4669 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.722519 4669 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.722406 4669 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.53:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186d1637aac4d5d1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-10 09:10:57.715770833 +0000 UTC m=+0.731789615,LastTimestamp:2025-10-10 09:10:57.715770833 +0000 UTC m=+0.731789615,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.724686 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.724842 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.725489 4669 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.725655 4669 factory.go:55] Registering systemd factory Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.725762 4669 factory.go:221] Registration of the systemd container factory successfully Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.726206 4669 factory.go:153] Registering CRI-O factory Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.726331 4669 factory.go:221] Registration of the crio container factory successfully Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.726446 4669 factory.go:103] Registering Raw factory Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.726527 4669 manager.go:1196] Started watching for new ooms in manager Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.727177 4669 server.go:460] "Adding debug handlers to kubelet server" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.727835 4669 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="200ms" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.733869 4669 manager.go:319] Starting recovery of all containers Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.741263 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.741440 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.741527 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.741700 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.741791 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.741872 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.741980 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.742059 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.742142 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.742229 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.742309 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.742391 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.743742 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744005 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744160 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744269 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744381 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744507 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744645 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744766 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744854 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.744935 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745014 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745098 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745211 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745325 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745425 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745538 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745696 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745788 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745887 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.745983 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.746064 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.746143 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.746222 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.746324 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.748542 4669 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.748691 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.748835 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.748923 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749016 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749099 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749178 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749258 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749335 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749413 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749624 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749714 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749810 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749902 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.749984 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750072 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750152 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750238 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750319 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750401 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750494 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750574 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750687 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750772 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750851 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.750965 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751060 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751140 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751244 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751347 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751446 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751545 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751668 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751754 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751832 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.751910 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752013 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752095 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752177 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752257 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752339 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752420 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752506 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752703 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752797 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752885 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.752965 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753061 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753148 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753229 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753325 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753405 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753495 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753573 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753680 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753761 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753842 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.753921 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754010 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754091 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754173 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754252 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754331 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754417 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754500 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754644 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754734 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754817 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754915 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.755004 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.755091 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.755190 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.755330 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.755485 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.755709 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.754767 4669 manager.go:324] Recovery completed Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.755829 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.756151 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.756236 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.757691 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.758030 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.758131 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.758212 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.758290 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.758382 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.758470 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759048 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759694 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759757 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759776 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759791 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759806 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759820 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759834 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759846 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759864 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759877 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759891 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759911 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759925 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759937 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759953 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759964 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759978 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.759991 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760004 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760016 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760030 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760050 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760069 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760086 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760102 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760115 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760131 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760144 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760155 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760168 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760181 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760192 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760205 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760216 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760260 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760273 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760289 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760306 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760325 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760338 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760348 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760358 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760371 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760382 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760395 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760410 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760429 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760445 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760461 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760475 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760489 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760503 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760517 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760530 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760546 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760560 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760573 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760588 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760616 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760630 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760643 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760657 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760672 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760685 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760698 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760712 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760725 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760738 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760750 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760762 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760775 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760787 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760798 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760810 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760823 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760834 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760845 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760859 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760871 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760883 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760894 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760905 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760918 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760929 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760941 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760954 4669 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760967 4669 reconstruct.go:97] "Volume reconstruction finished" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.760978 4669 reconciler.go:26] "Reconciler: start to sync state" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.768198 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.772294 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.772477 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.772576 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.778285 4669 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.778312 4669 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.778341 4669 state_mem.go:36] "Initialized new in-memory state store" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.790715 4669 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.791014 4669 policy_none.go:49] "None policy: Start" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.793836 4669 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.793947 4669 state_mem.go:35] "Initializing new in-memory state store" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.793906 4669 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.794124 4669 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.794161 4669 kubelet.go:2335] "Starting kubelet main sync loop" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.794217 4669 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 10 09:10:57 crc kubenswrapper[4669]: W1010 09:10:57.795242 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.795346 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.822103 4669 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.853052 4669 manager.go:334] "Starting Device Plugin manager" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.853107 4669 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.853120 4669 server.go:79] "Starting device plugin registration server" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.853497 4669 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.853514 4669 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.855963 4669 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.856090 4669 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.856104 4669 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.860506 4669 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.894570 4669 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.894720 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.897277 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.897306 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.897318 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.897472 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.897855 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.897926 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898196 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898227 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898242 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898339 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898462 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898495 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898894 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.898922 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899362 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899383 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899364 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899393 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899420 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899501 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899613 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.899640 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.900225 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.900243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.900251 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.900323 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.900512 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.900536 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901435 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901453 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901465 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901489 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901508 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901519 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901616 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901627 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901635 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901772 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.901791 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.902365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.902474 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.902561 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.929260 4669 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="400ms" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.956839 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.958693 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.958723 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.958735 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.958760 4669 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 09:10:57 crc kubenswrapper[4669]: E1010 09:10:57.959245 4669 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.965830 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.965869 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.965895 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.965920 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.965955 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.965987 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966033 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966061 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966085 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966101 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966128 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966146 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966184 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966237 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:57 crc kubenswrapper[4669]: I1010 09:10:57.966263 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067193 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067254 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067283 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067311 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067338 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067364 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067369 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067369 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067408 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067456 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067500 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067545 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067500 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067519 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067664 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067704 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067728 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067755 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067784 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067813 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067839 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067837 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067866 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067899 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067897 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067908 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067945 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067960 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.067975 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.068076 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.159756 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.161086 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.161135 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.161152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.161179 4669 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 09:10:58 crc kubenswrapper[4669]: E1010 09:10:58.161642 4669 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.241220 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.250376 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.264047 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.273892 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.280112 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.305818 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0f20327bf1c603b54b65c3e8120db6aec0581dd0f36b57873f18063da4c94435 WatchSource:0}: Error finding container 0f20327bf1c603b54b65c3e8120db6aec0581dd0f36b57873f18063da4c94435: Status 404 returned error can't find the container with id 0f20327bf1c603b54b65c3e8120db6aec0581dd0f36b57873f18063da4c94435 Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.313972 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-3c2a08e9716414df3b09692f0dac21d0a809c9bb041cb2e56c4eee84323cbfce WatchSource:0}: Error finding container 3c2a08e9716414df3b09692f0dac21d0a809c9bb041cb2e56c4eee84323cbfce: Status 404 returned error can't find the container with id 3c2a08e9716414df3b09692f0dac21d0a809c9bb041cb2e56c4eee84323cbfce Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.325830 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-9b6d215efb9b4c86d132fe3bd3b75800f60e14c234cd7d4845b3cb00badac186 WatchSource:0}: Error finding container 9b6d215efb9b4c86d132fe3bd3b75800f60e14c234cd7d4845b3cb00badac186: Status 404 returned error can't find the container with id 9b6d215efb9b4c86d132fe3bd3b75800f60e14c234cd7d4845b3cb00badac186 Oct 10 09:10:58 crc kubenswrapper[4669]: E1010 09:10:58.330619 4669 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="800ms" Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.339277 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-255259a3458320f5f81232cfb1e1278db7dd8ef2f0ebe7f8da3476e0261cd09d WatchSource:0}: Error finding container 255259a3458320f5f81232cfb1e1278db7dd8ef2f0ebe7f8da3476e0261cd09d: Status 404 returned error can't find the container with id 255259a3458320f5f81232cfb1e1278db7dd8ef2f0ebe7f8da3476e0261cd09d Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.562725 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.563968 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.564000 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.564012 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.564035 4669 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 09:10:58 crc kubenswrapper[4669]: E1010 09:10:58.564419 4669 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.663643 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:58 crc kubenswrapper[4669]: E1010 09:10:58.663740 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.717489 4669 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.721346 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:58 crc kubenswrapper[4669]: E1010 09:10:58.721452 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.771410 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:58 crc kubenswrapper[4669]: E1010 09:10:58.771513 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.800776 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"255259a3458320f5f81232cfb1e1278db7dd8ef2f0ebe7f8da3476e0261cd09d"} Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.802175 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9b6d215efb9b4c86d132fe3bd3b75800f60e14c234cd7d4845b3cb00badac186"} Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.803461 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3c2a08e9716414df3b09692f0dac21d0a809c9bb041cb2e56c4eee84323cbfce"} Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.805318 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7193d8a677a7d461b22af6f8910257d9abc8fa321aea52b8a7a303a26c217730"} Oct 10 09:10:58 crc kubenswrapper[4669]: I1010 09:10:58.806374 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f20327bf1c603b54b65c3e8120db6aec0581dd0f36b57873f18063da4c94435"} Oct 10 09:10:58 crc kubenswrapper[4669]: W1010 09:10:58.880201 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:58 crc kubenswrapper[4669]: E1010 09:10:58.880280 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:10:59 crc kubenswrapper[4669]: E1010 09:10:59.131457 4669 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="1.6s" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.364859 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.366721 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.366781 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.366801 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.366837 4669 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 09:10:59 crc kubenswrapper[4669]: E1010 09:10:59.367354 4669 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.717565 4669 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.811973 4669 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef" exitCode=0 Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.812093 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef"} Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.812152 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.813415 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.813460 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.813473 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.814935 4669 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964" exitCode=0 Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.815002 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.814983 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964"} Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.816232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.816284 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.816301 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.817543 4669 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f" exitCode=0 Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.817645 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.817642 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f"} Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.818636 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.818663 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.818675 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.823524 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940"} Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.823574 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d"} Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.823613 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07"} Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.824851 4669 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc" exitCode=0 Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.824892 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc"} Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.825012 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.825836 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.825858 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.825869 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.828439 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.830192 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.830243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:10:59 crc kubenswrapper[4669]: I1010 09:10:59.830257 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:00 crc kubenswrapper[4669]: W1010 09:11:00.515355 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:11:00 crc kubenswrapper[4669]: E1010 09:11:00.515434 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.716843 4669 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:11:00 crc kubenswrapper[4669]: E1010 09:11:00.732776 4669 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.53:6443: connect: connection refused" interval="3.2s" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.834526 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.834630 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.835718 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.835757 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.835770 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.839274 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.839304 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.839315 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.839429 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.841115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.841139 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.841149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.849632 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.849795 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.851194 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.851232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.851247 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.853136 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.853166 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.853184 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.853197 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.855026 4669 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5" exitCode=0 Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.855063 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5"} Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.855188 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.856030 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.856054 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.856066 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.967497 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.969312 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.969365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.969376 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:00 crc kubenswrapper[4669]: I1010 09:11:00.969404 4669 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 09:11:00 crc kubenswrapper[4669]: E1010 09:11:00.970002 4669 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.53:6443: connect: connection refused" node="crc" Oct 10 09:11:01 crc kubenswrapper[4669]: W1010 09:11:01.164952 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:11:01 crc kubenswrapper[4669]: E1010 09:11:01.165045 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:11:01 crc kubenswrapper[4669]: W1010 09:11:01.315553 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:11:01 crc kubenswrapper[4669]: E1010 09:11:01.315707 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:11:01 crc kubenswrapper[4669]: W1010 09:11:01.345386 4669 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.53:6443: connect: connection refused Oct 10 09:11:01 crc kubenswrapper[4669]: E1010 09:11:01.345533 4669 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.53:6443: connect: connection refused" logger="UnhandledError" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.860462 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.863127 4669 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="40ed6bd13a5887342822ef72ebe78bdea49922972d30b95d3563dccac17663e0" exitCode=255 Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.863227 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"40ed6bd13a5887342822ef72ebe78bdea49922972d30b95d3563dccac17663e0"} Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.863637 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.865160 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.865213 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.865223 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.865691 4669 scope.go:117] "RemoveContainer" containerID="40ed6bd13a5887342822ef72ebe78bdea49922972d30b95d3563dccac17663e0" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.870141 4669 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec" exitCode=0 Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.870284 4669 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.870337 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.871226 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.871950 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec"} Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.872142 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.872893 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.874145 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.874184 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.874205 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.874262 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.874332 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.874354 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.876101 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.876136 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.876147 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.876108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.876264 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:01 crc kubenswrapper[4669]: I1010 09:11:01.876287 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.190549 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.874227 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.876695 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02"} Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.876811 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.878018 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.878056 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.878067 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.880907 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.881318 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"766207b5eb0a9a74e33f9401e4e7973100af2d4fa385cef24239061db0e81ae8"} Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.881346 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c08061d946fd75010d0a38811215a27587ead13f95b1fd145a4011e19d8f96f1"} Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.881357 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d1db2b1fc6bf595859b800a987b85701bd3e814c627c1795a573ac49b8fd621e"} Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.881365 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f30092148564b7d90704c6cb8df802f66670871f12b3fac0ade6a238144ea204"} Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.881645 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.881665 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:02 crc kubenswrapper[4669]: I1010 09:11:02.881672 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.888668 4669 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.888679 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d6f5cb9f2e8b5efefffb5a810fc2bf904ef0ae855b723207b4d0b2b024ce6571"} Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.888724 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.888820 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.889842 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.889902 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.889925 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.890251 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.890316 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:03 crc kubenswrapper[4669]: I1010 09:11:03.890329 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.170987 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.172111 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.172140 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.172150 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.172169 4669 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.411407 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.555745 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.556015 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.558122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.558216 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.558239 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.559233 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.891644 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.893164 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.893210 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:04 crc kubenswrapper[4669]: I1010 09:11:04.893227 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.050158 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.050431 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.052046 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.052130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.052149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.055482 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.707113 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.893544 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.893544 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.895024 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.895065 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.895084 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.895312 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.895342 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:05 crc kubenswrapper[4669]: I1010 09:11:05.895353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.611227 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.611528 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.613366 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.613465 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.613487 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.896204 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.897890 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.897953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:06 crc kubenswrapper[4669]: I1010 09:11:06.897981 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.124293 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.124539 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.126193 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.126289 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.126316 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.133651 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:07 crc kubenswrapper[4669]: E1010 09:11:07.860738 4669 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.898795 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.899773 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.899800 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:07 crc kubenswrapper[4669]: I1010 09:11:07.899808 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.407171 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.407361 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.408984 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.409015 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.409023 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.412316 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.902616 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.903553 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.903605 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:09 crc kubenswrapper[4669]: I1010 09:11:09.903617 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:11 crc kubenswrapper[4669]: I1010 09:11:11.718725 4669 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 10 09:11:11 crc kubenswrapper[4669]: I1010 09:11:11.983237 4669 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 10 09:11:11 crc kubenswrapper[4669]: I1010 09:11:11.983358 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 10 09:11:11 crc kubenswrapper[4669]: I1010 09:11:11.991046 4669 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 10 09:11:11 crc kubenswrapper[4669]: I1010 09:11:11.991139 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 10 09:11:12 crc kubenswrapper[4669]: I1010 09:11:12.408995 4669 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 09:11:12 crc kubenswrapper[4669]: I1010 09:11:12.409113 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 09:11:13 crc kubenswrapper[4669]: I1010 09:11:13.454683 4669 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 09:11:13 crc kubenswrapper[4669]: I1010 09:11:13.454883 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.484881 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.485071 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.486789 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.486842 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.486853 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.501479 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.915863 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.917106 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.917230 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:14 crc kubenswrapper[4669]: I1010 09:11:14.917304 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.051463 4669 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.051562 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.062887 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.063090 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.063427 4669 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.063492 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.064565 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.064639 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.064660 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.070082 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.918548 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.919408 4669 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.919489 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.924011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.924050 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:15 crc kubenswrapper[4669]: I1010 09:11:15.924062 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:16 crc kubenswrapper[4669]: E1010 09:11:16.932507 4669 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.934967 4669 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.935098 4669 trace.go:236] Trace[1605980845]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 09:11:05.239) (total time: 11695ms): Oct 10 09:11:16 crc kubenswrapper[4669]: Trace[1605980845]: ---"Objects listed" error: 11695ms (09:11:16.935) Oct 10 09:11:16 crc kubenswrapper[4669]: Trace[1605980845]: [11.695102799s] [11.695102799s] END Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.935133 4669 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.937903 4669 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.938278 4669 trace.go:236] Trace[1673039121]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 09:11:05.010) (total time: 11928ms): Oct 10 09:11:16 crc kubenswrapper[4669]: Trace[1673039121]: ---"Objects listed" error: 11927ms (09:11:16.938) Oct 10 09:11:16 crc kubenswrapper[4669]: Trace[1673039121]: [11.928088751s] [11.928088751s] END Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.938300 4669 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 10 09:11:16 crc kubenswrapper[4669]: E1010 09:11:16.940448 4669 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.942858 4669 trace.go:236] Trace[1484816098]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 09:11:06.899) (total time: 10043ms): Oct 10 09:11:16 crc kubenswrapper[4669]: Trace[1484816098]: ---"Objects listed" error: 10043ms (09:11:16.942) Oct 10 09:11:16 crc kubenswrapper[4669]: Trace[1484816098]: [10.043435324s] [10.043435324s] END Oct 10 09:11:16 crc kubenswrapper[4669]: I1010 09:11:16.943057 4669 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.711796 4669 apiserver.go:52] "Watching apiserver" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.714293 4669 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.714574 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-922g5","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.714911 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.715019 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.715162 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.715239 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.715246 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.715288 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.715312 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.715375 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.715390 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.715672 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.717108 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.717310 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.718161 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.718478 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.718546 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.719057 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.720977 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.721144 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.721203 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.721601 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.722704 4669 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.724469 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.724547 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742842 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742875 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742894 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742912 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742931 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742947 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742962 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742979 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.742997 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743010 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743024 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743043 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743058 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743072 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743086 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743100 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743118 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743136 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743154 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743168 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743196 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743213 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743210 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743230 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743264 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743299 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743322 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743338 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743358 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743377 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743396 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743411 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743420 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743428 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743460 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743480 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743495 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743516 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743531 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743547 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743566 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743599 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743616 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743631 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743648 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743665 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743719 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743735 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743749 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743765 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743782 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743800 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743816 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743833 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743852 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743867 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743882 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743899 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743915 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743931 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743947 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743964 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743979 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743994 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744010 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744027 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744042 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743566 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744056 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743771 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744043 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.743896 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744150 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744251 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744299 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744422 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744567 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744598 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744630 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744741 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744756 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744904 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744977 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745148 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745246 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745411 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745417 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745567 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745695 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745795 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745854 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745901 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.745964 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.746135 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.746218 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.746444 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.750958 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751041 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751128 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751246 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751263 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751443 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751494 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751696 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.751794 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.752200 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.752243 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.752439 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.752521 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.752626 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.753140 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.753219 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.753263 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.754403 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.754932 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755158 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755399 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755653 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.744057 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755736 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755764 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755772 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755809 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755845 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755872 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755898 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.755933 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.756402 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.757030 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.757054 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.757390 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.758266 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.759270 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.761432 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.761984 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765197 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765262 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765289 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765336 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765365 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765409 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765437 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765465 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765511 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765538 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765576 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765637 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765661 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765696 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765713 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765731 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765766 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.765791 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.767476 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.767838 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.767941 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.768114 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.768488 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.769029 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.769305 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.769704 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.769731 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.769960 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.770166 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.770456 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.770528 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.770894 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.771237 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.771557 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.771831 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772158 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772337 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772366 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772398 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772421 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772466 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772490 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772507 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772543 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772615 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772638 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772654 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772671 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772704 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772721 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772735 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772770 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772787 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772804 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772822 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772859 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772885 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772930 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772950 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.772973 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773006 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773053 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773087 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773109 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773127 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773352 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773649 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773802 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773968 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.773989 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.774148 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.774238 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.774498 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.774691 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.774782 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.774914 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.774993 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.775053 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.775207 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.775276 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.775374 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.775522 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.775656 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.775951 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776028 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776375 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776378 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776401 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776448 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776487 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776512 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776534 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776551 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776570 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776603 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.776958 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.777105 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.777535 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.777603 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.777883 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.778336 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781040 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781114 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781137 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781162 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781180 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781200 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781222 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781242 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781263 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781281 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781298 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781315 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781330 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781350 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781368 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781385 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781402 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781420 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781438 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781453 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781470 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781489 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781507 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781526 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781546 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781606 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781623 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781646 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781665 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781685 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781700 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781715 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781730 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781748 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781766 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781785 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781805 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781824 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781842 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781858 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781875 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781892 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781907 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781924 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781938 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781956 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781972 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.781990 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782022 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782038 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782054 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782167 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782184 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782198 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782215 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782232 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782249 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782270 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782292 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782308 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782326 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782342 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782359 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782394 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782414 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782430 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782446 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782462 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782477 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782493 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782511 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782527 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782544 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.782580 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783266 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783301 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783322 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783354 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783433 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783458 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783480 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783491 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783499 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8t7n\" (UniqueName: \"kubernetes.io/projected/4486ac07-abf2-46bf-b5d3-354bde8d3b10-kube-api-access-b8t7n\") pod \"node-resolver-922g5\" (UID: \"4486ac07-abf2-46bf-b5d3-354bde8d3b10\") " pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783609 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783635 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783680 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783779 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783809 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783810 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783857 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4486ac07-abf2-46bf-b5d3-354bde8d3b10-hosts-file\") pod \"node-resolver-922g5\" (UID: \"4486ac07-abf2-46bf-b5d3-354bde8d3b10\") " pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783887 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783941 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.783985 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784151 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784243 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784269 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784314 4669 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784327 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784339 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784352 4669 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784389 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784399 4669 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784409 4669 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784419 4669 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784420 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784428 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784468 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784484 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784500 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784513 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784531 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784576 4669 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784635 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784647 4669 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784659 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784670 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784682 4669 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784693 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784706 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784717 4669 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784728 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784738 4669 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784749 4669 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784760 4669 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784770 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784781 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784777 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784792 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784844 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784865 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784881 4669 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.784900 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785018 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785034 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785045 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785056 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785066 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785076 4669 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785085 4669 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785095 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785106 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785116 4669 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785126 4669 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785136 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785145 4669 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785155 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785164 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785173 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785184 4669 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785193 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785202 4669 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785211 4669 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785221 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785230 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785239 4669 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785250 4669 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785259 4669 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785269 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785278 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785287 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785298 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785308 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785317 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785326 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785336 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785348 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785357 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785368 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785377 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785385 4669 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785395 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785405 4669 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785416 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785427 4669 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785440 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785447 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785451 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785530 4669 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785540 4669 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785551 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785681 4669 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785692 4669 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785703 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785713 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785725 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785768 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785778 4669 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785788 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785799 4669 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785809 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785908 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785919 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785931 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785943 4669 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.785952 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786171 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786614 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786627 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786637 4669 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786648 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786657 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786667 4669 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786835 4669 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786845 4669 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786856 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786866 4669 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786874 4669 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786974 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786987 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.786996 4669 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.787005 4669 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.787015 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.787874 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.787928 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:18.287912047 +0000 UTC m=+21.303930789 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.790407 4669 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.790711 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.793268 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.796083 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.796185 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:18.296164202 +0000 UTC m=+21.312182944 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.797178 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.797403 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.797573 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:11:18.297559494 +0000 UTC m=+21.313578236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.797698 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.797851 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.798148 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.798669 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.799241 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.799374 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.799551 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.799706 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.799869 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.800023 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.800429 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.801326 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.810791 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.811215 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.811523 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.811782 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.812277 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.812567 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.812602 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.812623 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.812823 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:18.312796954 +0000 UTC m=+21.328815906 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.812911 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.814359 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.816502 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.816698 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.816825 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.816838 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.816851 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.816901 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:18.31688356 +0000 UTC m=+21.332902492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.817558 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.817860 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.819533 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.819762 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.820127 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.822816 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.823056 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.823572 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.823701 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.823812 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.823882 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.823969 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.823986 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.824243 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.825148 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.825166 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.825338 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.825349 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.825528 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.825689 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826044 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826072 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826052 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826266 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826295 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826302 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826321 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826329 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826500 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826573 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826714 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826725 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.826903 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.830884 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.831916 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.832390 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.834091 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.827534 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.834982 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.835400 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.835837 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.836019 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.837327 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.838049 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.838057 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.838093 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.838649 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.838777 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.838920 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.839098 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.839823 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.840057 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.840517 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.840657 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.840741 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.841645 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.842428 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.842637 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.847417 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.848398 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.848896 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.850707 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.850815 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.851655 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.856048 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.865731 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.871361 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.872335 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.869481 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.877645 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.878439 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.878508 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.878694 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.879938 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.880759 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.881562 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.882814 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.883321 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.886773 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.887564 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.888338 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889209 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4486ac07-abf2-46bf-b5d3-354bde8d3b10-hosts-file\") pod \"node-resolver-922g5\" (UID: \"4486ac07-abf2-46bf-b5d3-354bde8d3b10\") " pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889358 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889521 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889561 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8t7n\" (UniqueName: \"kubernetes.io/projected/4486ac07-abf2-46bf-b5d3-354bde8d3b10-kube-api-access-b8t7n\") pod \"node-resolver-922g5\" (UID: \"4486ac07-abf2-46bf-b5d3-354bde8d3b10\") " pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889634 4669 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889645 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889654 4669 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889664 4669 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889675 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889686 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889696 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889706 4669 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889715 4669 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889724 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889734 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889744 4669 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889753 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889764 4669 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889773 4669 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889782 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889792 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889803 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889812 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889821 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889830 4669 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889840 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889849 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889859 4669 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889870 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889880 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889890 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889899 4669 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889908 4669 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889958 4669 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.889986 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4486ac07-abf2-46bf-b5d3-354bde8d3b10-hosts-file\") pod \"node-resolver-922g5\" (UID: \"4486ac07-abf2-46bf-b5d3-354bde8d3b10\") " pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890174 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890654 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890673 4669 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890684 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890693 4669 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890702 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890712 4669 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890720 4669 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890729 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890738 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890747 4669 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890755 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890764 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890775 4669 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890783 4669 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890792 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890802 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890811 4669 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890820 4669 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890830 4669 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890838 4669 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890847 4669 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890857 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890866 4669 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890874 4669 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890884 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890893 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890905 4669 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890916 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890927 4669 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890939 4669 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890950 4669 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890962 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.890972 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891071 4669 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891084 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891096 4669 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891107 4669 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891116 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891125 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891134 4669 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891142 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891185 4669 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891195 4669 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891203 4669 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891211 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891220 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891237 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891251 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891264 4669 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891307 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.891460 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.892167 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.892557 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.892672 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.893258 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.897556 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.901849 4669 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.902127 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.906534 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.907187 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.908575 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.910281 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.914121 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.915050 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.915909 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8t7n\" (UniqueName: \"kubernetes.io/projected/4486ac07-abf2-46bf-b5d3-354bde8d3b10-kube-api-access-b8t7n\") pod \"node-resolver-922g5\" (UID: \"4486ac07-abf2-46bf-b5d3-354bde8d3b10\") " pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.916335 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.917163 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.918814 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.920278 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.920513 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.923338 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.924569 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.925166 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.926011 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.926558 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.926892 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.927448 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.927651 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.928152 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.929044 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.929554 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.929600 4669 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02" exitCode=255 Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.930012 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.930888 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.931501 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.932328 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.932772 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02"} Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.932838 4669 scope.go:117] "RemoveContainer" containerID="40ed6bd13a5887342822ef72ebe78bdea49922972d30b95d3563dccac17663e0" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.934989 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.948486 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.949283 4669 scope.go:117] "RemoveContainer" containerID="91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.949770 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 10 09:11:17 crc kubenswrapper[4669]: E1010 09:11:17.949804 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.958133 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.966088 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.976556 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.985355 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.992251 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:11:17 crc kubenswrapper[4669]: I1010 09:11:17.994920 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.006306 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40ed6bd13a5887342822ef72ebe78bdea49922972d30b95d3563dccac17663e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"message\\\":\\\"W1010 09:11:01.066354 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1010 09:11:01.066777 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760087461 cert, and key in /tmp/serving-cert-3432688085/serving-signer.crt, /tmp/serving-cert-3432688085/serving-signer.key\\\\nI1010 09:11:01.368793 1 observer_polling.go:159] Starting file observer\\\\nW1010 09:11:01.372731 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1010 09:11:01.373001 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:01.374456 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3432688085/tls.crt::/tmp/serving-cert-3432688085/tls.key\\\\\\\"\\\\nF1010 09:11:01.694068 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.019062 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.028839 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.031922 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.042398 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.050608 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.062178 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.073896 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-922g5" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.076922 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.090643 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 09:11:18 crc kubenswrapper[4669]: W1010 09:11:18.103434 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4486ac07_abf2_46bf_b5d3_354bde8d3b10.slice/crio-8216ccbf69426d6398838a6017c2c28ae1585db916fef0fd05f98e7a8c6d409e WatchSource:0}: Error finding container 8216ccbf69426d6398838a6017c2c28ae1585db916fef0fd05f98e7a8c6d409e: Status 404 returned error can't find the container with id 8216ccbf69426d6398838a6017c2c28ae1585db916fef0fd05f98e7a8c6d409e Oct 10 09:11:18 crc kubenswrapper[4669]: W1010 09:11:18.113160 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-e901a4d8600679798ba05075c8085b81fd36cbcc7db3491bd6895d2a36d56e03 WatchSource:0}: Error finding container e901a4d8600679798ba05075c8085b81fd36cbcc7db3491bd6895d2a36d56e03: Status 404 returned error can't find the container with id e901a4d8600679798ba05075c8085b81fd36cbcc7db3491bd6895d2a36d56e03 Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.300901 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.300968 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.301023 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.301084 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.301125 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:19.301113755 +0000 UTC m=+22.317132497 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.301170 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:11:19.301162837 +0000 UTC m=+22.317181579 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.301217 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.301234 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:19.301229489 +0000 UTC m=+22.317248231 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.401751 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.401802 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.401924 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.401938 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.401959 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.401965 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.402002 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.402014 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.402004 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:19.401989505 +0000 UTC m=+22.418008247 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.402081 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:19.402066447 +0000 UTC m=+22.418085179 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.934198 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.936224 4669 scope.go:117] "RemoveContainer" containerID="91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02" Oct 10 09:11:18 crc kubenswrapper[4669]: E1010 09:11:18.936355 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.937668 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.937707 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.937718 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e901a4d8600679798ba05075c8085b81fd36cbcc7db3491bd6895d2a36d56e03"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.938541 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"edd12c369f20192e9aa44f140450202028245eeb75f317f0a7e89466314b333d"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.939852 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.939894 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"675b89251122471d48d225e8c79659748473de2bad83e0616bd8060eab677361"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.940771 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-922g5" event={"ID":"4486ac07-abf2-46bf-b5d3-354bde8d3b10","Type":"ContainerStarted","Data":"0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.940800 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-922g5" event={"ID":"4486ac07-abf2-46bf-b5d3-354bde8d3b10","Type":"ContainerStarted","Data":"8216ccbf69426d6398838a6017c2c28ae1585db916fef0fd05f98e7a8c6d409e"} Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.947711 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.957003 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.966473 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.979052 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:18 crc kubenswrapper[4669]: I1010 09:11:18.992753 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.012116 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.037373 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.053287 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.074087 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.090855 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.103683 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.115118 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.126038 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.137407 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.148286 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.160308 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.310887 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.310940 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.310971 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.311080 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.311127 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:21.311114117 +0000 UTC m=+24.327132859 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.311405 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:11:21.311397706 +0000 UTC m=+24.327416448 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.311437 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.311459 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:21.311453148 +0000 UTC m=+24.327471880 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.410941 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.412110 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.412147 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412244 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412257 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412266 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412301 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:21.412290316 +0000 UTC m=+24.428309058 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412551 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412599 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412611 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.412675 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:21.412651037 +0000 UTC m=+24.428669779 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.417256 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.422683 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.433252 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.445411 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.457943 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.469737 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-nq6jn"] Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.470276 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.471107 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vbjz8"] Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.471857 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.473253 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.473280 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.473290 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.474223 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.475113 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.476423 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.480326 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.499871 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.504652 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.516184 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.527281 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.539680 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.552526 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.565645 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.578599 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.594420 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.607699 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614298 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-etc-kubernetes\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614358 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-os-release\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614385 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-hostroot\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614406 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-multus-certs\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614430 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6583f591-656e-4230-9772-b2cd0d5176c0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614449 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9502d677-e546-4df7-96c2-bce8274c0f57-cni-binary-copy\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-system-cni-dir\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614487 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-cni-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614567 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9502d677-e546-4df7-96c2-bce8274c0f57-multus-daemon-config\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614634 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-os-release\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614669 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-k8s-cni-cncf-io\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614697 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-cni-multus\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614723 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-cnibin\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614744 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614782 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-system-cni-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614806 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-cnibin\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614826 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-socket-dir-parent\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614847 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-kubelet\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614867 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-conf-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614905 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spxq9\" (UniqueName: \"kubernetes.io/projected/6583f591-656e-4230-9772-b2cd0d5176c0-kube-api-access-spxq9\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614921 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6583f591-656e-4230-9772-b2cd0d5176c0-cni-binary-copy\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614937 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-netns\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614960 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-cni-bin\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.614977 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm2ph\" (UniqueName: \"kubernetes.io/projected/9502d677-e546-4df7-96c2-bce8274c0f57-kube-api-access-fm2ph\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.616527 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.628352 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.639178 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.655198 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.669455 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.690120 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715656 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-netns\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715691 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-cni-bin\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715706 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm2ph\" (UniqueName: \"kubernetes.io/projected/9502d677-e546-4df7-96c2-bce8274c0f57-kube-api-access-fm2ph\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715720 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-hostroot\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715736 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-multus-certs\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715750 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-etc-kubernetes\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715772 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-os-release\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715786 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9502d677-e546-4df7-96c2-bce8274c0f57-cni-binary-copy\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715800 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6583f591-656e-4230-9772-b2cd0d5176c0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715818 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-system-cni-dir\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715832 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-cni-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715854 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9502d677-e546-4df7-96c2-bce8274c0f57-multus-daemon-config\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715854 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-etc-kubernetes\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715876 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-os-release\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715872 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-hostroot\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715903 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-k8s-cni-cncf-io\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715922 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-cni-multus\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715937 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-cnibin\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715953 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715967 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-system-cni-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715981 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-cnibin\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715994 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-socket-dir-parent\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716009 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-kubelet\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716023 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-conf-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716028 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-os-release\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716045 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6583f591-656e-4230-9772-b2cd0d5176c0-cni-binary-copy\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716060 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spxq9\" (UniqueName: \"kubernetes.io/projected/6583f591-656e-4230-9772-b2cd0d5176c0-kube-api-access-spxq9\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716208 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-multus-certs\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716244 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-cnibin\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716214 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-cni-multus\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716288 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-k8s-cni-cncf-io\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716321 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-kubelet\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716327 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-system-cni-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716355 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-conf-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716368 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-cni-dir\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716377 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-cnibin\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716392 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-system-cni-dir\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716421 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-multus-socket-dir-parent\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716428 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-os-release\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.715824 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-var-lib-cni-bin\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716564 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9502d677-e546-4df7-96c2-bce8274c0f57-cni-binary-copy\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716664 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9502d677-e546-4df7-96c2-bce8274c0f57-host-run-netns\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716663 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6583f591-656e-4230-9772-b2cd0d5176c0-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716806 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6583f591-656e-4230-9772-b2cd0d5176c0-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.716947 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6583f591-656e-4230-9772-b2cd0d5176c0-cni-binary-copy\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.717114 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9502d677-e546-4df7-96c2-bce8274c0f57-multus-daemon-config\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.732186 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm2ph\" (UniqueName: \"kubernetes.io/projected/9502d677-e546-4df7-96c2-bce8274c0f57-kube-api-access-fm2ph\") pod \"multus-nq6jn\" (UID: \"9502d677-e546-4df7-96c2-bce8274c0f57\") " pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.734857 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spxq9\" (UniqueName: \"kubernetes.io/projected/6583f591-656e-4230-9772-b2cd0d5176c0-kube-api-access-spxq9\") pod \"multus-additional-cni-plugins-vbjz8\" (UID: \"6583f591-656e-4230-9772-b2cd0d5176c0\") " pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.785714 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nq6jn" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.797065 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.797241 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.797576 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.797707 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.797802 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:19 crc kubenswrapper[4669]: E1010 09:11:19.797908 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.798281 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.800385 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.801049 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.802844 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.803476 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.804072 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 10 09:11:19 crc kubenswrapper[4669]: W1010 09:11:19.831785 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6583f591_656e_4230_9772_b2cd0d5176c0.slice/crio-16d52c46380211fa5df376ce86eba057618c6fb3905d0662f056a2c7fb561c90 WatchSource:0}: Error finding container 16d52c46380211fa5df376ce86eba057618c6fb3905d0662f056a2c7fb561c90: Status 404 returned error can't find the container with id 16d52c46380211fa5df376ce86eba057618c6fb3905d0662f056a2c7fb561c90 Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.889124 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-x6v7p"] Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.889639 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.900967 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.901061 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.912055 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.912278 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.912477 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.920225 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gbsxj"] Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.920993 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.926166 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.930672 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.930775 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.930837 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.930844 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.930880 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.930881 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.947958 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerStarted","Data":"745a15c72b811bf0d9b9a4b2aba48ff560a7cf96c16d66bd725dcb2b8755ee44"} Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.961963 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.986619 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerStarted","Data":"16d52c46380211fa5df376ce86eba057618c6fb3905d0662f056a2c7fb561c90"} Oct 10 09:11:19 crc kubenswrapper[4669]: I1010 09:11:19.995835 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:19Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020722 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-config\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020756 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/addb758f-1f34-4793-af67-1a54167543b9-mcd-auth-proxy-config\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020776 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-ovn-kubernetes\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020792 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020809 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-systemd\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020823 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-etc-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020892 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-script-lib\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020942 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020972 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-ovn\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.020993 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-env-overrides\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021021 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-bin\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021042 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovn-node-metrics-cert\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021086 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/addb758f-1f34-4793-af67-1a54167543b9-proxy-tls\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021107 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-node-log\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021126 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-log-socket\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021150 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/addb758f-1f34-4793-af67-1a54167543b9-rootfs\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021170 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqx62\" (UniqueName: \"kubernetes.io/projected/addb758f-1f34-4793-af67-1a54167543b9-kube-api-access-bqx62\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021216 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-systemd-units\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021240 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-netns\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021260 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-slash\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021282 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-netd\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021305 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-var-lib-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021325 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xsrl\" (UniqueName: \"kubernetes.io/projected/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-kube-api-access-6xsrl\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.021345 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-kubelet\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.043345 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.068290 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.093868 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122045 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-kubelet\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122307 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-var-lib-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-var-lib-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122184 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-kubelet\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122405 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xsrl\" (UniqueName: \"kubernetes.io/projected/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-kube-api-access-6xsrl\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122488 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-config\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122512 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/addb758f-1f34-4793-af67-1a54167543b9-mcd-auth-proxy-config\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122528 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-ovn-kubernetes\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122547 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122565 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-systemd\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122619 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-ovn-kubernetes\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122634 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122678 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-etc-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122696 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-script-lib\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122728 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-systemd\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122731 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-etc-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122762 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-env-overrides\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122781 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122824 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-openvswitch\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.122976 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-ovn\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123192 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-env-overrides\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123230 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-ovn\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123350 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-config\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123540 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-script-lib\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123544 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/addb758f-1f34-4793-af67-1a54167543b9-mcd-auth-proxy-config\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123574 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovn-node-metrics-cert\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123642 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-bin\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123661 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/addb758f-1f34-4793-af67-1a54167543b9-proxy-tls\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123676 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-node-log\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123698 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-log-socket\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123723 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/addb758f-1f34-4793-af67-1a54167543b9-rootfs\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123722 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-bin\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123763 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-node-log\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123784 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-log-socket\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123803 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/addb758f-1f34-4793-af67-1a54167543b9-rootfs\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123737 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqx62\" (UniqueName: \"kubernetes.io/projected/addb758f-1f34-4793-af67-1a54167543b9-kube-api-access-bqx62\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123835 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-systemd-units\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123854 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-slash\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123869 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-netns\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123887 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-netd\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123920 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-netd\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123942 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-netns\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123947 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-slash\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.123963 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-systemd-units\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.128902 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.129248 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovn-node-metrics-cert\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.129496 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/addb758f-1f34-4793-af67-1a54167543b9-proxy-tls\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.148708 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.162101 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqx62\" (UniqueName: \"kubernetes.io/projected/addb758f-1f34-4793-af67-1a54167543b9-kube-api-access-bqx62\") pod \"machine-config-daemon-x6v7p\" (UID: \"addb758f-1f34-4793-af67-1a54167543b9\") " pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.170602 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xsrl\" (UniqueName: \"kubernetes.io/projected/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-kube-api-access-6xsrl\") pod \"ovnkube-node-gbsxj\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.173555 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.190477 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.204097 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.207081 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.224294 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.240548 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.255661 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.280032 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.305483 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.322941 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.340123 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.357817 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.372550 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.386732 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.405476 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.418171 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.429445 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.439618 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.454636 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.493290 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:20Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.990408 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b"} Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.991853 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerStarted","Data":"79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620"} Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.993313 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c" exitCode=0 Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.993360 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.993376 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"732d505dfba05c3478bc1aba7ec5bed96b5acf93eabbdceef4e9c76b0307bba5"} Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.995469 4669 generic.go:334] "Generic (PLEG): container finished" podID="6583f591-656e-4230-9772-b2cd0d5176c0" containerID="ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754" exitCode=0 Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.995610 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerDied","Data":"ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754"} Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.998257 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576"} Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.998306 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c"} Oct 10 09:11:20 crc kubenswrapper[4669]: I1010 09:11:20.998323 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"f6e60f2e3d7723a54cc59650bb1e4cd67a7e2c944f1ad131e83783a5bbb6ffca"} Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.019107 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.036223 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.052982 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.073640 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.093263 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.106012 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.147798 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.177567 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.201062 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.236285 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.254657 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.272075 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.302222 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.324346 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.334553 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.334694 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:11:25.334670271 +0000 UTC m=+28.350689013 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.334739 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.334854 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.334873 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.334908 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:25.334892648 +0000 UTC m=+28.350911390 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.334943 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.334973 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:25.3349672 +0000 UTC m=+28.350985942 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.340694 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.355126 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.368225 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.379795 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.401144 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.416513 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.430399 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.435517 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.435565 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435710 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435741 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435710 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435754 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435769 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435780 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435815 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:25.435799518 +0000 UTC m=+28.451818270 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.435831 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:25.435824279 +0000 UTC m=+28.451843021 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.446138 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.456758 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.468554 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.480444 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.504361 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:21Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.797064 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.797310 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:21 crc kubenswrapper[4669]: I1010 09:11:21.797362 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.797382 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.797451 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:21 crc kubenswrapper[4669]: E1010 09:11:21.797508 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.005296 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.005342 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.005354 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.005365 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.005377 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.005387 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.007269 4669 generic.go:334] "Generic (PLEG): container finished" podID="6583f591-656e-4230-9772-b2cd0d5176c0" containerID="dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f" exitCode=0 Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.007298 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerDied","Data":"dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f"} Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.021585 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.035821 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.061805 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.083918 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.097751 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.113420 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.128008 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.136315 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.149328 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.160817 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.176326 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-pvhp4"] Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.176408 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.177395 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.179337 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.179337 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.179467 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.179553 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.192388 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.205878 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.218464 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.229936 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.251291 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.266815 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.279895 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.294137 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.307277 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.318487 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.334295 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.344854 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3bb764f3-806c-4614-b6a4-f247ff3fa796-serviceca\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.344896 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktp2\" (UniqueName: \"kubernetes.io/projected/3bb764f3-806c-4614-b6a4-f247ff3fa796-kube-api-access-tktp2\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.344976 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3bb764f3-806c-4614-b6a4-f247ff3fa796-host\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.352563 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.365108 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.377884 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.389842 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.401578 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.445989 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3bb764f3-806c-4614-b6a4-f247ff3fa796-serviceca\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.446024 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktp2\" (UniqueName: \"kubernetes.io/projected/3bb764f3-806c-4614-b6a4-f247ff3fa796-kube-api-access-tktp2\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.446053 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3bb764f3-806c-4614-b6a4-f247ff3fa796-host\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.446114 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3bb764f3-806c-4614-b6a4-f247ff3fa796-host\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.447196 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3bb764f3-806c-4614-b6a4-f247ff3fa796-serviceca\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.471813 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktp2\" (UniqueName: \"kubernetes.io/projected/3bb764f3-806c-4614-b6a4-f247ff3fa796-kube-api-access-tktp2\") pod \"node-ca-pvhp4\" (UID: \"3bb764f3-806c-4614-b6a4-f247ff3fa796\") " pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: I1010 09:11:22.599202 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pvhp4" Oct 10 09:11:22 crc kubenswrapper[4669]: W1010 09:11:22.613399 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bb764f3_806c_4614_b6a4_f247ff3fa796.slice/crio-27105e5e064881b5b73d2acd4314033b2c52fa9b661cc8451be4fa2aaed0ded2 WatchSource:0}: Error finding container 27105e5e064881b5b73d2acd4314033b2c52fa9b661cc8451be4fa2aaed0ded2: Status 404 returned error can't find the container with id 27105e5e064881b5b73d2acd4314033b2c52fa9b661cc8451be4fa2aaed0ded2 Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.013709 4669 generic.go:334] "Generic (PLEG): container finished" podID="6583f591-656e-4230-9772-b2cd0d5176c0" containerID="3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5" exitCode=0 Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.013760 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerDied","Data":"3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.016295 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pvhp4" event={"ID":"3bb764f3-806c-4614-b6a4-f247ff3fa796","Type":"ContainerStarted","Data":"16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.016352 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pvhp4" event={"ID":"3bb764f3-806c-4614-b6a4-f247ff3fa796","Type":"ContainerStarted","Data":"27105e5e064881b5b73d2acd4314033b2c52fa9b661cc8451be4fa2aaed0ded2"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.028897 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.048530 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.060031 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.071550 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.084752 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.097992 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.112207 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.125453 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.140399 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.149853 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.159159 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.169605 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.182074 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.199358 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.211891 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.221706 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.232627 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.243192 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.254614 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.266223 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.280166 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.292935 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.306270 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.315078 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.323830 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.334308 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.340914 4669 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.342468 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.342500 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.342531 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.342933 4669 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.346225 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.348218 4669 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.348441 4669 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.349240 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.349265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.349274 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.349286 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.349295 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.364798 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.367473 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.367507 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.367518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.367533 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.367543 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.376497 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.379949 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.383156 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.383191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.383204 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.383221 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.383233 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.402723 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.405728 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.405769 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.405787 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.405808 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.405823 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.426028 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.429004 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.429047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.429059 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.429081 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.429096 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.440069 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.440233 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.441370 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.441407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.441482 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.441500 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.441510 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.453939 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.454609 4669 scope.go:117] "RemoveContainer" containerID="91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02" Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.454792 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.543304 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.543347 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.543359 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.543375 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.543387 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.646477 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.646519 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.646531 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.646553 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.646566 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.748070 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.748094 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.748103 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.748115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.748123 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.795644 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.795644 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.795769 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.795806 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.796134 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:23 crc kubenswrapper[4669]: E1010 09:11:23.796356 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.850480 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.850741 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.850807 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.850874 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.850932 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.954541 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.954762 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.954839 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.954958 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:23 crc kubenswrapper[4669]: I1010 09:11:23.955040 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:23Z","lastTransitionTime":"2025-10-10T09:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.025134 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.027742 4669 generic.go:334] "Generic (PLEG): container finished" podID="6583f591-656e-4230-9772-b2cd0d5176c0" containerID="17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0" exitCode=0 Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.027806 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerDied","Data":"17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.057521 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.057557 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.057568 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.057625 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.057638 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.062139 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.080435 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.098468 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.111469 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.128498 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.151439 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.159829 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.159866 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.159877 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.159893 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.159904 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.169502 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.182759 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.199858 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.210900 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.219914 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.231054 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.253991 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.262439 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.262470 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.262480 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.262494 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.262504 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.273615 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.365831 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.365877 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.365887 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.365899 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.365908 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.469151 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.469191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.469201 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.469217 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.469227 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.571200 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.571233 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.571242 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.571256 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.571266 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.673563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.673621 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.673631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.673645 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.673655 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.775255 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.775291 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.775300 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.775313 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.775322 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.877211 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.877252 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.877265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.877282 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.877294 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.979713 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.979758 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.979772 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.979791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:24 crc kubenswrapper[4669]: I1010 09:11:24.979804 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:24Z","lastTransitionTime":"2025-10-10T09:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.034864 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerStarted","Data":"6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.082294 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.082332 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.082340 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.082354 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.082363 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.184497 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.184538 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.184553 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.184572 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.184616 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.287833 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.287881 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.287898 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.287916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.287932 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.375232 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.375372 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.375410 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.375528 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.375647 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:33.375629007 +0000 UTC m=+36.391647749 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.375710 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:11:33.375701799 +0000 UTC m=+36.391720541 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.375742 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.375763 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:33.375757661 +0000 UTC m=+36.391776403 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.390118 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.390159 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.390170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.390188 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.390199 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.476186 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.476252 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476346 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476365 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476375 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476375 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476401 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476417 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476420 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:33.476407674 +0000 UTC m=+36.492426416 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.476480 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:33.476463716 +0000 UTC m=+36.492482488 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.493335 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.493370 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.493383 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.493398 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.493409 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.596808 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.596867 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.596884 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.596906 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.596956 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.699178 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.699224 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.699238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.699258 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.699274 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.795355 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.795406 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.795534 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.795619 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.795816 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:25 crc kubenswrapper[4669]: E1010 09:11:25.795993 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.801916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.801979 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.802002 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.802030 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.802052 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.905098 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.905156 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.905176 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.905206 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:25 crc kubenswrapper[4669]: I1010 09:11:25.905227 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:25Z","lastTransitionTime":"2025-10-10T09:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.008039 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.008104 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.008122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.008145 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.008163 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.061374 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.084475 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.097294 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.110940 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.110999 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.111016 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.111041 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.111060 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.118212 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.136405 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.156712 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.175117 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.195903 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.213385 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.214126 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.214191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.214218 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.214249 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.214271 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.225039 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.238194 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.258538 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.275959 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.289893 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:26Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.317335 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.317391 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.317407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.317427 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.317441 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.419321 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.419540 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.419659 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.419779 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.420105 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.522980 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.523048 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.523072 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.523128 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.523152 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.625160 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.625238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.625263 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.625292 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.625311 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.728298 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.728358 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.728375 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.728399 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.728415 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.830670 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.830714 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.830727 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.830744 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.830757 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.933568 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.933724 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.933744 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.933770 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:26 crc kubenswrapper[4669]: I1010 09:11:26.933790 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:26Z","lastTransitionTime":"2025-10-10T09:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.036622 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.036688 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.036711 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.036739 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.036760 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.139384 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.139420 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.139433 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.139449 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.139461 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.242000 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.242044 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.242054 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.242071 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.242082 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.345000 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.345170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.345190 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.345210 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.345224 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.449004 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.449101 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.449534 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.449881 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.450016 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.553372 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.553420 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.553437 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.553487 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.553504 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.656968 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.657025 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.657047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.657076 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.657100 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.759706 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.759792 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.759809 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.759832 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.759848 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.795465 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:27 crc kubenswrapper[4669]: E1010 09:11:27.795639 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.795945 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:27 crc kubenswrapper[4669]: E1010 09:11:27.796046 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.796210 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:27 crc kubenswrapper[4669]: E1010 09:11:27.796366 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.819474 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.839978 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.857733 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.864635 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.864680 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.864693 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.864708 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.864718 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.883508 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.898641 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.911709 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.927283 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.953485 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.965495 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.967177 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.967220 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.967234 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.967254 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.967265 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:27Z","lastTransitionTime":"2025-10-10T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.977815 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.988821 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:27 crc kubenswrapper[4669]: I1010 09:11:27.996182 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.007079 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:28Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.018663 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:28Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.070041 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.070085 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.070095 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.070113 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.070126 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.173901 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.173955 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.173984 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.174024 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.174053 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.276381 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.276431 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.276449 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.276471 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.276489 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.379536 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.379567 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.379575 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.379603 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.379612 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.482421 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.482493 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.482505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.482521 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.482863 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.585195 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.585244 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.585261 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.585287 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.585305 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.688434 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.688514 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.688538 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.688567 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.688636 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.791036 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.791064 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.791071 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.791084 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.791093 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.894302 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.894340 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.894354 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.894375 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.894387 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.997310 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.997346 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.997358 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.997375 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:28 crc kubenswrapper[4669]: I1010 09:11:28.997392 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:28Z","lastTransitionTime":"2025-10-10T09:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.057758 4669 generic.go:334] "Generic (PLEG): container finished" podID="6583f591-656e-4230-9772-b2cd0d5176c0" containerID="6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18" exitCode=0 Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.057816 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerDied","Data":"6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.098218 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.099908 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.099947 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.099959 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.099975 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.099986 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.150084 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.182304 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.193222 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.201566 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.201612 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.201621 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.201635 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.201646 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.208941 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.219968 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.230818 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.242250 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.260130 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.271291 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.284367 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.297059 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.304191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.304221 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.304229 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.304247 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.304256 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.312098 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.324160 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:29Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.410374 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.410405 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.410414 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.410429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.410437 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.512653 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.512688 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.512696 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.512712 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.512720 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.614654 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.614716 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.614734 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.614758 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.614776 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.717126 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.718483 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.718502 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.718520 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.718540 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.795205 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.795266 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.795207 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:29 crc kubenswrapper[4669]: E1010 09:11:29.795333 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:29 crc kubenswrapper[4669]: E1010 09:11:29.795430 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:29 crc kubenswrapper[4669]: E1010 09:11:29.795471 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.821273 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.821315 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.821329 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.821347 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.821361 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.924773 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.924866 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.924891 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.924918 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:29 crc kubenswrapper[4669]: I1010 09:11:29.924942 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:29Z","lastTransitionTime":"2025-10-10T09:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.028244 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.028297 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.028313 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.028337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.028354 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.131696 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.131814 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.131841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.131914 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.131996 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.234774 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.234852 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.234876 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.234903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.234920 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.338307 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.338373 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.338391 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.338414 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.338431 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.441755 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.441799 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.441815 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.441837 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.441853 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.544932 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.545011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.545030 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.545057 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.545075 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.648406 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.648460 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.648479 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.648505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.648522 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.751935 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.751977 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.751988 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.752007 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.752022 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.855149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.855252 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.855273 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.855301 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.855318 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.958137 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.958197 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.958218 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.958297 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:30 crc kubenswrapper[4669]: I1010 09:11:30.958318 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:30Z","lastTransitionTime":"2025-10-10T09:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.060707 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.060748 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.060758 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.060774 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.060787 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.068800 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.163344 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.163406 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.163421 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.163442 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.163454 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.266675 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.266712 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.266726 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.266743 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.266756 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.369327 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.369361 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.369372 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.369387 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.369398 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.472455 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.472518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.472540 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.472569 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.472628 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.576193 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.576229 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.576238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.576251 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.576261 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.678171 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.678213 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.678228 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.678243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.678254 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.780736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.780858 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.780992 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.781018 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.781033 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.795218 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:31 crc kubenswrapper[4669]: E1010 09:11:31.795356 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.795444 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.795546 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:31 crc kubenswrapper[4669]: E1010 09:11:31.795576 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:31 crc kubenswrapper[4669]: E1010 09:11:31.795934 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.883919 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.883978 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.883992 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.884017 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.884032 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.986953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.987019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.987033 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.987086 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:31 crc kubenswrapper[4669]: I1010 09:11:31.987101 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:31Z","lastTransitionTime":"2025-10-10T09:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.018913 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p"] Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.019614 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.023893 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.024336 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.044606 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.050953 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.051204 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.051435 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d4st\" (UniqueName: \"kubernetes.io/projected/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-kube-api-access-2d4st\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.051566 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.062299 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.074764 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerStarted","Data":"a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.083647 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.088982 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.089034 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.089043 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.089056 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.089066 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.097428 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.107754 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.122207 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.133616 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.145108 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.152987 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.153696 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d4st\" (UniqueName: \"kubernetes.io/projected/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-kube-api-access-2d4st\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.153756 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.153822 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.154513 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.154552 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.157669 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.167751 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.170946 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d4st\" (UniqueName: \"kubernetes.io/projected/eaa7fbe4-1c87-4c6b-befc-64a1473973b3-kube-api-access-2d4st\") pod \"ovnkube-control-plane-749d76644c-2v44p\" (UID: \"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.178836 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.189690 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.191649 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.191703 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.191717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.191730 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.191739 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.204369 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.216496 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.225459 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.235109 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:32Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.293962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.293991 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.294001 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.294015 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.294025 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.331924 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.397952 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.397998 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.398013 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.398043 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.398059 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.500555 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.500644 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.500657 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.500674 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.500685 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.602943 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.603003 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.603019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.603048 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.603065 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.705107 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.705130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.705138 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.705150 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.705158 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.813219 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.813265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.813277 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.813295 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.813305 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.916388 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.916442 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.916461 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.916486 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:32 crc kubenswrapper[4669]: I1010 09:11:32.916504 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:32Z","lastTransitionTime":"2025-10-10T09:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.019671 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.019725 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.019737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.019757 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.019770 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.079315 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" event={"ID":"eaa7fbe4-1c87-4c6b-befc-64a1473973b3","Type":"ContainerStarted","Data":"9726268aee2a475f349188bf791d6dbce4f741299eec2bd014f5b34e9f0ee02d"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.083145 4669 generic.go:334] "Generic (PLEG): container finished" podID="6583f591-656e-4230-9772-b2cd0d5176c0" containerID="a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8" exitCode=0 Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.083192 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerDied","Data":"a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.083740 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.083951 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.084021 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.109687 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.122345 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.122377 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.122386 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.122413 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.122424 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.123842 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.171784 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.173657 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.175101 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.177931 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-bhn82"] Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.178301 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.178356 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.198008 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.215203 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.224896 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.224942 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.224959 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.224990 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.225007 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.227526 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.240654 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.255559 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.266254 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.266333 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc9bg\" (UniqueName: \"kubernetes.io/projected/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-kube-api-access-nc9bg\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.271026 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.294464 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.312229 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.330843 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.332911 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.332952 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.332961 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.332978 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.332987 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.345917 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.362623 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.366899 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.366977 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc9bg\" (UniqueName: \"kubernetes.io/projected/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-kube-api-access-nc9bg\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.367339 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.367404 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:11:33.867383433 +0000 UTC m=+36.883402195 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.380308 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.391362 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.392376 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc9bg\" (UniqueName: \"kubernetes.io/projected/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-kube-api-access-nc9bg\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.406282 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.417664 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.429735 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.435758 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.435785 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.435794 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.435811 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.435822 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.442749 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.456179 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.465260 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.467304 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.467431 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.467476 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:11:49.467455107 +0000 UTC m=+52.483473879 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.467536 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.467563 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.467578 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:49.467569451 +0000 UTC m=+52.483588193 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.467642 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.467692 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:49.467679324 +0000 UTC m=+52.483698146 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.477569 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.491787 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.506454 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.518332 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.530978 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.538166 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.538199 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.538211 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.538227 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.538238 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.543682 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.554765 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.568352 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.568404 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568518 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568542 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568554 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568518 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568615 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:49.568599525 +0000 UTC m=+52.584618267 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568626 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568638 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.568676 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:11:49.568663757 +0000 UTC m=+52.584682499 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.571605 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.582986 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.640084 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.640115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.640125 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.640138 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.640148 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.742500 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.742556 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.742573 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.742613 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.742625 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.794819 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.794893 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.794916 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.794946 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.794981 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.795102 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.823232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.823280 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.823295 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.823314 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.823330 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.840160 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.844031 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.844055 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.844063 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.844075 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.844085 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.857241 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.865185 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.865461 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.865637 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.865801 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.865923 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.870841 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.870956 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.870991 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:11:34.870979315 +0000 UTC m=+37.886998057 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.883437 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.887083 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.887134 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.887150 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.887501 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.887546 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.901389 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.905015 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.905045 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.905059 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.905080 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.905094 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.920795 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:33Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:33 crc kubenswrapper[4669]: E1010 09:11:33.920933 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.922841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.922974 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.923076 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.923180 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:33 crc kubenswrapper[4669]: I1010 09:11:33.923274 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:33Z","lastTransitionTime":"2025-10-10T09:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.026749 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.027026 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.027148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.027238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.027331 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.090221 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" event={"ID":"eaa7fbe4-1c87-4c6b-befc-64a1473973b3","Type":"ContainerStarted","Data":"2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.090278 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" event={"ID":"eaa7fbe4-1c87-4c6b-befc-64a1473973b3","Type":"ContainerStarted","Data":"74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.096033 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" event={"ID":"6583f591-656e-4230-9772-b2cd0d5176c0","Type":"ContainerStarted","Data":"e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.116098 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.129634 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.129664 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.129675 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.129690 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.129700 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.131537 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.142994 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.158895 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.183916 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.197901 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.212369 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.228393 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.233237 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.233417 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.233506 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.233575 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.233670 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.249692 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.265002 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.286895 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.300121 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.317550 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.333154 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.337248 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.337313 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.337328 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.337354 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.337368 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.347153 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.360383 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:34Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.441031 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.441104 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.441121 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.441152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.441173 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.545124 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.545459 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.545569 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.545737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.545830 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.648364 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.648405 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.648418 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.648444 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.648460 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.752037 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.752318 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.752394 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.752429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.752498 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.795167 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:34 crc kubenswrapper[4669]: E1010 09:11:34.795386 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.855032 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.855072 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.855080 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.855094 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.855103 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.880847 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:34 crc kubenswrapper[4669]: E1010 09:11:34.880992 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:34 crc kubenswrapper[4669]: E1010 09:11:34.881049 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:11:36.881034969 +0000 UTC m=+39.897053711 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.957835 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.957870 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.957880 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.957896 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:34 crc kubenswrapper[4669]: I1010 09:11:34.957908 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:34Z","lastTransitionTime":"2025-10-10T09:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.060052 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.060112 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.060129 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.060158 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.060178 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.120640 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.133128 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" probeResult="failure" output="" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.140454 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.152503 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.162656 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.162690 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.162700 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.162744 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.162755 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.166623 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.179920 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.192136 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.202033 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.216442 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.228620 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.240656 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.249776 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.260923 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.264748 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.264773 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.264782 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.264794 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.264802 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.274680 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.291328 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.303078 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.316029 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.368533 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.368573 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.368669 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.368782 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.368807 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.471811 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.471882 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.471902 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.471926 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.471943 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.574085 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.574115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.574125 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.574140 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.574151 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.675918 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.675952 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.675964 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.675979 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.675989 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.779192 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.779239 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.779246 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.779259 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.779270 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.795229 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.795260 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:35 crc kubenswrapper[4669]: E1010 09:11:35.795362 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.795389 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:35 crc kubenswrapper[4669]: E1010 09:11:35.795501 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:35 crc kubenswrapper[4669]: E1010 09:11:35.795610 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.885972 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.886009 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.886018 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.886034 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.886044 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.988617 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.988668 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.988684 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.988703 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:35 crc kubenswrapper[4669]: I1010 09:11:35.988717 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:35Z","lastTransitionTime":"2025-10-10T09:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.091414 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.091447 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.091458 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.091473 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.091486 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.193277 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.193340 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.193353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.193370 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.193383 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.295813 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.295848 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.295859 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.295905 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.295920 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.398907 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.398944 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.398953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.398967 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.398977 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.501557 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.501617 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.501629 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.501644 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.501655 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.605107 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.605148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.605160 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.605176 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.605187 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.707781 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.707814 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.707825 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.707840 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.707853 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.794821 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:36 crc kubenswrapper[4669]: E1010 09:11:36.795021 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.811647 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.811702 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.811717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.811736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.811748 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.905042 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:36 crc kubenswrapper[4669]: E1010 09:11:36.905216 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:36 crc kubenswrapper[4669]: E1010 09:11:36.905271 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:11:40.905254342 +0000 UTC m=+43.921273094 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.914773 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.914810 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.914824 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.914841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:36 crc kubenswrapper[4669]: I1010 09:11:36.914853 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:36Z","lastTransitionTime":"2025-10-10T09:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.016916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.016976 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.016995 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.017018 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.017034 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.107440 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/0.log" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.109934 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272" exitCode=1 Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.109972 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.110697 4669 scope.go:117] "RemoveContainer" containerID="39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.120218 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.120258 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.120268 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.120283 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.120292 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.128072 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.148121 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.176175 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:36Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 09:11:36.457304 5885 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:36.457342 5885 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 09:11:36.457372 5885 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 09:11:36.457381 5885 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:36.457419 5885 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 09:11:36.457475 5885 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 09:11:36.457499 5885 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 09:11:36.457517 5885 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 09:11:36.457529 5885 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 09:11:36.457535 5885 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 09:11:36.457548 5885 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 09:11:36.457558 5885 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 09:11:36.457570 5885 factory.go:656] Stopping watch factory\\\\nI1010 09:11:36.457577 5885 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 09:11:36.457599 5885 ovnkube.go:599] Stopped ovnkube\\\\nI1010 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.187072 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.202167 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.217271 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.222118 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.222149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.222158 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.222173 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.222184 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.239729 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.282618 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.298862 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.317407 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.324802 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.324834 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.324843 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.324858 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.324866 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.332721 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.341731 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.352901 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.364038 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.379073 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.391034 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.427424 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.427510 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.427527 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.427552 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.427569 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.530563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.530662 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.530680 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.530706 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.530727 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.634212 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.634271 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.634291 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.634315 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.634331 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.736523 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.736563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.736573 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.736618 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.736630 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.794768 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.794797 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:37 crc kubenswrapper[4669]: E1010 09:11:37.794896 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.794961 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:37 crc kubenswrapper[4669]: E1010 09:11:37.795251 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:37 crc kubenswrapper[4669]: E1010 09:11:37.795374 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.795489 4669 scope.go:117] "RemoveContainer" containerID="91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.808708 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.830176 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.843424 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.843461 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.843471 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.843486 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.843497 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.853125 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.865078 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.885401 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.903658 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.920251 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.932486 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.943787 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.946279 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.946315 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.946328 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.946351 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.946365 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:37Z","lastTransitionTime":"2025-10-10T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.958715 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.975943 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:37 crc kubenswrapper[4669]: I1010 09:11:37.992279 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.008562 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.023420 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.041529 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.048706 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.048729 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.048737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.048750 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.048758 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.059222 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:36Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 09:11:36.457304 5885 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:36.457342 5885 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 09:11:36.457372 5885 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 09:11:36.457381 5885 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:36.457419 5885 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 09:11:36.457475 5885 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 09:11:36.457499 5885 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 09:11:36.457517 5885 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 09:11:36.457529 5885 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 09:11:36.457535 5885 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 09:11:36.457548 5885 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 09:11:36.457558 5885 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 09:11:36.457570 5885 factory.go:656] Stopping watch factory\\\\nI1010 09:11:36.457577 5885 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 09:11:36.457599 5885 ovnkube.go:599] Stopped ovnkube\\\\nI1010 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.116575 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/0.log" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.118653 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.119088 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.129513 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.141026 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.151040 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.151077 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.151088 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.151104 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.151114 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.153058 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.166250 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.176081 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.185882 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.197015 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.213623 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:36Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 09:11:36.457304 5885 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:36.457342 5885 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 09:11:36.457372 5885 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 09:11:36.457381 5885 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:36.457419 5885 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 09:11:36.457475 5885 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 09:11:36.457499 5885 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 09:11:36.457517 5885 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 09:11:36.457529 5885 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 09:11:36.457535 5885 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 09:11:36.457548 5885 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 09:11:36.457558 5885 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 09:11:36.457570 5885 factory.go:656] Stopping watch factory\\\\nI1010 09:11:36.457577 5885 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 09:11:36.457599 5885 ovnkube.go:599] Stopped ovnkube\\\\nI1010 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.223978 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.234144 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.244472 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.255447 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.255501 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.255517 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.255540 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.255554 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.257558 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.270059 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.283749 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.296088 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.305010 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.357795 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.357841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.357853 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.357873 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.357887 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.460686 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.460730 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.460743 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.460760 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.460772 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.563744 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.563771 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.563779 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.563791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.563799 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.666197 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.666234 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.666243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.666258 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.666269 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.768240 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.768281 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.768292 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.768309 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.768321 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.794756 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:38 crc kubenswrapper[4669]: E1010 09:11:38.794909 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.871236 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.871283 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.871295 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.871315 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.871327 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.974202 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.974259 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.974274 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.974295 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:38 crc kubenswrapper[4669]: I1010 09:11:38.974311 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:38Z","lastTransitionTime":"2025-10-10T09:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.077096 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.077172 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.077195 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.077223 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.077245 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.124022 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/1.log" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.124878 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/0.log" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.128108 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056" exitCode=1 Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.128153 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.128184 4669 scope.go:117] "RemoveContainer" containerID="39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.128773 4669 scope.go:117] "RemoveContainer" containerID="7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056" Oct 10 09:11:39 crc kubenswrapper[4669]: E1010 09:11:39.128975 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.139361 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.144374 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.144982 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.148186 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.160399 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.180024 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.180412 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.180511 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.180568 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.180661 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.180755 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.202130 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.218224 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.238430 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.257537 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.284037 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.284097 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.284114 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.284152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.284167 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.284628 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:36Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 09:11:36.457304 5885 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:36.457342 5885 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 09:11:36.457372 5885 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 09:11:36.457381 5885 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:36.457419 5885 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 09:11:36.457475 5885 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 09:11:36.457499 5885 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 09:11:36.457517 5885 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 09:11:36.457529 5885 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 09:11:36.457535 5885 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 09:11:36.457548 5885 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 09:11:36.457558 5885 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 09:11:36.457570 5885 factory.go:656] Stopping watch factory\\\\nI1010 09:11:36.457577 5885 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 09:11:36.457599 5885 ovnkube.go:599] Stopped ovnkube\\\\nI1010 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.298035 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.318454 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.336408 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.351780 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.370580 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.386185 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.388135 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.388173 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.388183 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.388199 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.388211 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.402255 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.415415 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.431053 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.447134 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.469431 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://39c929a2406eb15d77938c23ac671ae4c5fedc9ec62195a17a4b14a51866a272\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:36Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 09:11:36.457304 5885 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:36.457342 5885 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 09:11:36.457372 5885 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 09:11:36.457381 5885 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:36.457419 5885 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 09:11:36.457475 5885 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 09:11:36.457499 5885 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 09:11:36.457517 5885 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 09:11:36.457529 5885 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 09:11:36.457535 5885 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1010 09:11:36.457548 5885 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 09:11:36.457558 5885 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 09:11:36.457570 5885 factory.go:656] Stopping watch factory\\\\nI1010 09:11:36.457577 5885 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1010 09:11:36.457599 5885 ovnkube.go:599] Stopped ovnkube\\\\nI1010 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.481180 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.490919 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.490962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.490974 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.490992 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.491003 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.497573 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.515347 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.532996 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.547775 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.567419 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.582133 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.593077 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.593118 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.593130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.593148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.593161 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.595923 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.607128 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.618433 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.630686 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.643439 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.655401 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:39Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.698819 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.698863 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.698875 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.698893 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.698905 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.794916 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:39 crc kubenswrapper[4669]: E1010 09:11:39.795256 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.795010 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:39 crc kubenswrapper[4669]: E1010 09:11:39.795580 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.795009 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:39 crc kubenswrapper[4669]: E1010 09:11:39.795871 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.801776 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.801938 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.802058 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.802167 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.802246 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.904698 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.904761 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.904778 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.904801 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:39 crc kubenswrapper[4669]: I1010 09:11:39.904819 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:39Z","lastTransitionTime":"2025-10-10T09:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.008097 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.008152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.008169 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.008188 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.008203 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.111105 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.111163 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.111180 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.111203 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.111218 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.150247 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/1.log" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.154403 4669 scope.go:117] "RemoveContainer" containerID="7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056" Oct 10 09:11:40 crc kubenswrapper[4669]: E1010 09:11:40.154637 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.171204 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.185698 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.197207 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.207182 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.214014 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.214040 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.214048 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.214108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.214121 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.218169 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.234545 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.253214 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.267595 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.283330 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.297680 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.316871 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.316918 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.316934 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.316954 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.316971 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.318015 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.330720 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.345251 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.363824 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.379757 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.393519 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:40Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.419854 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.419928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.419947 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.419981 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.420004 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.523646 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.523708 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.523725 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.523751 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.523768 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.626255 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.626289 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.626299 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.626316 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.626326 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.728427 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.728461 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.728469 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.728481 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.728489 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.794748 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:40 crc kubenswrapper[4669]: E1010 09:11:40.794893 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.831495 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.831538 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.831549 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.831566 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.831613 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.934118 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.934174 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.934191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.934210 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.934221 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:40Z","lastTransitionTime":"2025-10-10T09:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:40 crc kubenswrapper[4669]: I1010 09:11:40.947756 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:40 crc kubenswrapper[4669]: E1010 09:11:40.948036 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:40 crc kubenswrapper[4669]: E1010 09:11:40.948102 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:11:48.948085478 +0000 UTC m=+51.964104220 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.037452 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.037521 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.037570 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.037633 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.037651 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.140908 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.140961 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.140978 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.140996 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.141008 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.244303 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.244739 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.244806 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.244872 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.244931 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.349159 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.349205 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.349221 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.349247 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.349264 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.452026 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.452100 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.452122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.452146 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.452163 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.555935 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.555993 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.556046 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.556074 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.556092 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.660567 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.660656 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.660676 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.660701 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.660720 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.763410 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.763475 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.763498 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.763526 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.763547 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.795275 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.795326 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.795289 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:41 crc kubenswrapper[4669]: E1010 09:11:41.795512 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:41 crc kubenswrapper[4669]: E1010 09:11:41.795653 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:41 crc kubenswrapper[4669]: E1010 09:11:41.795790 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.867051 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.867148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.867215 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.867243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.867297 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.970719 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.970787 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.970811 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.970839 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:41 crc kubenswrapper[4669]: I1010 09:11:41.970861 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:41Z","lastTransitionTime":"2025-10-10T09:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.074152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.074217 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.074237 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.074262 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.074280 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.177615 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.177671 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.177688 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.177712 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.177730 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.279812 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.279879 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.279900 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.279928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.279951 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.382932 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.382983 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.382995 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.383011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.383052 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.485625 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.485676 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.485688 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.485705 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.485718 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.587926 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.587978 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.587996 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.588019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.588036 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.690803 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.690858 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.690878 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.690902 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.690922 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.794464 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:42 crc kubenswrapper[4669]: E1010 09:11:42.794669 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.795089 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.795130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.795138 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.795154 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.795165 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.897460 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.897498 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.897510 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.897529 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:42 crc kubenswrapper[4669]: I1010 09:11:42.897544 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:42.999807 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:42.999846 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:42.999860 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:42.999877 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:42.999888 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:42Z","lastTransitionTime":"2025-10-10T09:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.102550 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.102650 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.102669 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.102693 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.102716 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.205802 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.206004 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.206086 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.206117 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.206184 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.309829 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.309913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.309943 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.309960 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.310011 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.413193 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.413247 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.413266 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.413292 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.413309 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.516645 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.517097 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.517243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.517386 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.517629 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.621279 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.621329 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.621343 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.621361 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.621374 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.725692 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.725747 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.725764 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.725790 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.725808 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.795903 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:43 crc kubenswrapper[4669]: E1010 09:11:43.796188 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.796349 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.796409 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:43 crc kubenswrapper[4669]: E1010 09:11:43.796646 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:43 crc kubenswrapper[4669]: E1010 09:11:43.796795 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.828973 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.829016 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.829029 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.829043 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.829055 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.931870 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.931930 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.931941 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.931958 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:43 crc kubenswrapper[4669]: I1010 09:11:43.931968 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:43Z","lastTransitionTime":"2025-10-10T09:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.035381 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.035716 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.035984 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.036136 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.036269 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.139261 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.139475 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.139572 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.139691 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.139761 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.181042 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.181158 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.181185 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.181214 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.181236 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: E1010 09:11:44.202294 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:44Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.207793 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.207884 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.207903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.207927 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.207946 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: E1010 09:11:44.226157 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:44Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.231337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.231391 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.231402 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.231448 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.231459 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: E1010 09:11:44.250486 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:44Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.256246 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.256313 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.256326 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.256346 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.256358 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: E1010 09:11:44.274075 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:44Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.278179 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.278225 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.278237 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.278259 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.278271 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: E1010 09:11:44.290303 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:44Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:44 crc kubenswrapper[4669]: E1010 09:11:44.290445 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.291832 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.291859 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.291870 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.291889 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.291903 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.394917 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.394977 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.394993 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.395016 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.395032 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.498305 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.498337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.498348 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.498363 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.498373 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.603329 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.603420 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.603440 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.603465 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.603483 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.705722 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.706352 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.706438 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.706523 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.706635 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.795025 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:44 crc kubenswrapper[4669]: E1010 09:11:44.795215 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.811066 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.811501 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.811679 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.811873 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.812050 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.915234 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.915303 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.915326 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.915357 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:44 crc kubenswrapper[4669]: I1010 09:11:44.915382 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:44Z","lastTransitionTime":"2025-10-10T09:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.019730 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.019796 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.019816 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.019843 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.019867 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.123028 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.123096 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.123120 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.123148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.123205 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.227018 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.227078 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.227095 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.227118 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.227134 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.330298 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.330366 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.330386 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.330411 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.330428 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.433683 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.433749 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.433769 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.433792 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.433810 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.537046 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.537105 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.537122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.537148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.537168 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.640394 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.640464 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.640487 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.640512 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.640530 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.744073 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.744161 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.744189 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.744224 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.744249 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.795393 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.795429 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:45 crc kubenswrapper[4669]: E1010 09:11:45.795553 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.795661 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:45 crc kubenswrapper[4669]: E1010 09:11:45.795778 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:45 crc kubenswrapper[4669]: E1010 09:11:45.795880 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.846494 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.846541 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.846552 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.846567 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.846579 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.949638 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.949679 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.949691 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.949709 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:45 crc kubenswrapper[4669]: I1010 09:11:45.949719 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:45Z","lastTransitionTime":"2025-10-10T09:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.052903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.052972 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.052993 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.053019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.053036 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.156209 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.156263 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.156276 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.156299 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.156311 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.259923 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.260031 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.260050 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.260077 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.260096 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.363727 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.363790 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.363814 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.363845 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.363868 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.467371 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.467468 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.467492 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.467520 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.467540 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.570663 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.570737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.570762 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.570791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.570811 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.674161 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.674258 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.674277 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.674311 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.674333 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.777556 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.777668 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.777691 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.777718 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.777737 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.795113 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:46 crc kubenswrapper[4669]: E1010 09:11:46.795314 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.880661 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.880733 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.880750 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.880778 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.880794 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.983980 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.984019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.984030 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.984044 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:46 crc kubenswrapper[4669]: I1010 09:11:46.984053 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:46Z","lastTransitionTime":"2025-10-10T09:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.087825 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.087887 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.087904 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.087928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.087944 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.191487 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.191545 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.191559 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.191580 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.191627 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.294808 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.294854 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.294862 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.294874 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.294883 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.397388 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.397443 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.397459 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.397481 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.397498 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.500311 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.500339 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.500349 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.500361 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.500369 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.604177 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.604247 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.604269 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.604297 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.604317 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.706657 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.706716 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.706741 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.706765 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.706785 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.794784 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.794932 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:47 crc kubenswrapper[4669]: E1010 09:11:47.795049 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.795111 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:47 crc kubenswrapper[4669]: E1010 09:11:47.795255 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:47 crc kubenswrapper[4669]: E1010 09:11:47.795466 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.809574 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.809624 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.809635 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.809649 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.809662 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.819333 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.841165 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.861003 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.876691 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.896407 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.914002 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.914096 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.914134 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.914156 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.914185 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.914208 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:47Z","lastTransitionTime":"2025-10-10T09:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.931723 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.961734 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.977235 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:47 crc kubenswrapper[4669]: I1010 09:11:47.993102 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:47Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.006679 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:48Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.017340 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.017370 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.017381 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.017397 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.017408 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.021400 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:48Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.034851 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:48Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.050731 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:48Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.083794 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:48Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.097380 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:48Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.123507 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.123558 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.123570 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.123616 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.123629 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.226315 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.226374 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.226399 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.226428 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.226449 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.333976 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.334056 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.334082 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.334112 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.334135 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.436619 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.436667 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.436685 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.436709 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.436728 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.540232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.540283 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.540307 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.540339 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.540360 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.642719 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.642780 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.642803 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.642831 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.642852 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.746561 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.746645 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.746669 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.746700 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.746720 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.794402 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:48 crc kubenswrapper[4669]: E1010 09:11:48.794685 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.850493 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.850563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.850618 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.850649 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.850672 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.953238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.953307 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.953329 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.953356 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:48 crc kubenswrapper[4669]: I1010 09:11:48.953377 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:48Z","lastTransitionTime":"2025-10-10T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.042805 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.043077 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.043181 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:12:05.043153386 +0000 UTC m=+68.059172158 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.058085 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.058214 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.058240 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.058769 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.058982 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.161713 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.161766 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.161784 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.161809 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.161826 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.264122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.264197 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.264219 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.264248 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.264271 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.366034 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.366085 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.366093 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.366109 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.366118 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.468737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.468779 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.468790 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.468805 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.468817 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.548679 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.548859 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.548885 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:12:21.548851044 +0000 UTC m=+84.564869796 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.548932 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.549089 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.549164 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.549229 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:12:21.549198524 +0000 UTC m=+84.565217316 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.549281 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:12:21.549262446 +0000 UTC m=+84.565281298 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.571377 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.571434 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.571447 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.571473 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.571488 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.650222 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.650342 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650425 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650493 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650507 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650582 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650635 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:12:21.65061468 +0000 UTC m=+84.666633442 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650655 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650680 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.650768 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:12:21.650739624 +0000 UTC m=+84.666758446 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.674498 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.674547 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.674565 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.674655 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.674669 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.778049 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.778078 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.778091 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.778107 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.778118 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.794724 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.794756 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.794794 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.794904 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.795036 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:49 crc kubenswrapper[4669]: E1010 09:11:49.795185 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.881852 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.881913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.881935 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.881963 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.881986 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.984135 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.984225 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.984239 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.984254 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:49 crc kubenswrapper[4669]: I1010 09:11:49.984266 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:49Z","lastTransitionTime":"2025-10-10T09:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.087369 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.087420 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.087437 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.087458 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.087472 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.190174 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.190254 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.190279 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.190308 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.190334 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.294492 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.294944 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.295249 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.295479 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.295798 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.398704 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.398934 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.399033 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.399159 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.399347 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.502975 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.503292 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.503513 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.503694 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.503832 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.607047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.607501 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.607729 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.607870 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.608001 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.711644 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.711689 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.711705 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.711731 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.711747 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.795094 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:50 crc kubenswrapper[4669]: E1010 09:11:50.795311 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.814249 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.814286 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.814302 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.814321 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.814334 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.916639 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.916672 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.916682 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.916698 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:50 crc kubenswrapper[4669]: I1010 09:11:50.916709 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:50Z","lastTransitionTime":"2025-10-10T09:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.019004 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.019035 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.019044 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.019057 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.019066 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.121616 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.121695 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.121717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.121740 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.121756 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.224180 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.224237 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.224250 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.224268 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.224279 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.328077 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.328155 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.328170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.328188 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.328201 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.431756 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.431846 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.431881 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.431916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.431937 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.535297 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.535344 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.535356 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.535377 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.535390 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.638429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.638498 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.638516 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.638541 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.638560 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.742371 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.742445 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.742471 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.742504 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.742530 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.794969 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.795143 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.794978 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:51 crc kubenswrapper[4669]: E1010 09:11:51.795238 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:51 crc kubenswrapper[4669]: E1010 09:11:51.795405 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:51 crc kubenswrapper[4669]: E1010 09:11:51.795524 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.845457 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.845554 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.845575 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.845658 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.845682 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.948927 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.948965 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.948973 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.948989 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:51 crc kubenswrapper[4669]: I1010 09:11:51.948998 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:51Z","lastTransitionTime":"2025-10-10T09:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.052171 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.052244 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.052269 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.052299 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.052326 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.155922 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.155975 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.155984 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.156002 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.156013 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.259747 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.259824 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.259837 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.259855 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.259866 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.363344 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.363391 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.363404 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.363424 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.363436 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.466557 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.466672 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.466686 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.466707 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.466720 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.571109 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.571176 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.571191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.571213 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.571236 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.674083 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.674230 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.674256 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.674281 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.674302 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.777987 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.778351 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.778372 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.778393 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.778410 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.794725 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:52 crc kubenswrapper[4669]: E1010 09:11:52.794985 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.884414 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.884876 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.885133 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.885453 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.885632 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.988942 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.989751 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.989776 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.989809 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:52 crc kubenswrapper[4669]: I1010 09:11:52.989825 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:52Z","lastTransitionTime":"2025-10-10T09:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.093707 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.093764 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.093783 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.093809 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.093828 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.196726 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.196778 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.196792 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.196815 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.196829 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.299555 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.299668 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.299686 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.299711 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.299729 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.401715 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.401802 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.401813 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.401826 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.401836 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.504283 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.504395 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.504406 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.504424 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.504436 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.606996 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.607046 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.607058 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.607077 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.607089 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.711663 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.711728 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.711746 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.711773 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.711792 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.795113 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.795222 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:53 crc kubenswrapper[4669]: E1010 09:11:53.795526 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.795554 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:53 crc kubenswrapper[4669]: E1010 09:11:53.796171 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:53 crc kubenswrapper[4669]: E1010 09:11:53.796347 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.796791 4669 scope.go:117] "RemoveContainer" containerID="7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.814088 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.814134 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.814151 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.814174 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.814194 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.917749 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.918235 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.918255 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.918283 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:53 crc kubenswrapper[4669]: I1010 09:11:53.918307 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:53Z","lastTransitionTime":"2025-10-10T09:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.021245 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.021298 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.021316 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.021340 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.021357 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.123665 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.123711 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.123726 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.123744 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.123758 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.206213 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/1.log" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.211071 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.213137 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.226131 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.226165 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.226174 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.226190 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.226202 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.226754 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.242657 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.269401 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.286368 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.307269 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.319800 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.328128 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.328157 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.328166 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.328179 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.328188 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.338608 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.359297 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.377186 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.388802 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.405137 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.418278 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.428052 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.430611 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.430632 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.430642 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.430654 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.430663 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.440370 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.451459 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.461540 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.532922 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.532960 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.532969 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.532986 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.532996 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.560776 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.572716 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.580192 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.593965 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.604011 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.616953 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.632532 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.635115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.635166 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.635325 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.635359 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.635370 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.644747 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.656674 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.668221 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.674394 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.674427 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.674438 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.674454 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.674463 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.679864 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: E1010 09:11:54.687261 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.690501 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.690544 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.690557 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.690578 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.690609 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.691072 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: E1010 09:11:54.703386 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.708757 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.708798 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.708812 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.708838 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.708854 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.713998 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: E1010 09:11:54.720020 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.729003 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.729045 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.729055 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.729075 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.729087 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.738563 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: E1010 09:11:54.743528 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.747058 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.747093 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.747104 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.747122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.747134 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.751629 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: E1010 09:11:54.760771 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: E1010 09:11:54.760879 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.762302 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.762329 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.762338 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.762353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.762363 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.764884 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.775049 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.786976 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:54Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.795247 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:54 crc kubenswrapper[4669]: E1010 09:11:54.795382 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.865271 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.865306 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.865315 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.865328 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.865338 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.968019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.968078 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.968094 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.968117 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:54 crc kubenswrapper[4669]: I1010 09:11:54.968135 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:54Z","lastTransitionTime":"2025-10-10T09:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.055803 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.070664 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.070726 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.070745 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.070776 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.070795 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.075616 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.089365 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.106395 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.118174 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.132977 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.145528 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.157091 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.170032 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.172935 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.172995 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.173011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.173033 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.173048 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.185137 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.201351 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.215945 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/2.log" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.216614 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/1.log" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.219248 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530" exitCode=1 Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.219305 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.219400 4669 scope.go:117] "RemoveContainer" containerID="7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.220567 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.220820 4669 scope.go:117] "RemoveContainer" containerID="ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530" Oct 10 09:11:55 crc kubenswrapper[4669]: E1010 09:11:55.221165 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.238798 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.250076 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.266881 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.275989 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.276036 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.276045 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.276064 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.276077 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.287220 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.308527 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.325026 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.339089 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.354874 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.370438 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.379484 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.379546 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.379563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.379610 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.379631 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.394157 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.405888 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.420838 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.434528 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.449659 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.462752 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.474402 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.482001 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.482047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.482081 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.482100 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.482111 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.490708 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.502034 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.512706 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.528079 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.542729 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.559508 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.583902 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7261c265f97b5a5cf3214f2ab719092a71ec39083fc7d94c572dd9915552f056\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:38Z\\\",\\\"message\\\":\\\"nformers/externalversions/factory.go:140\\\\nI1010 09:11:38.132769 6121 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 09:11:38.132790 6121 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.132824 6121 factory.go:656] Stopping watch factory\\\\nI1010 09:11:38.132842 6121 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 09:11:38.132920 6121 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.133042 6121 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1010 09:11:38.141020 6121 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1010 09:11:38.141070 6121 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1010 09:11:38.141169 6121 ovnkube.go:599] Stopped ovnkube\\\\nI1010 09:11:38.141213 6121 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 09:11:38.141390 6121 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:55Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.585099 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.585154 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.585167 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.585185 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.585197 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.688299 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.688443 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.688468 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.688495 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.688519 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.791522 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.791579 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.791627 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.791650 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.791666 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.795166 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.795191 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:55 crc kubenswrapper[4669]: E1010 09:11:55.795344 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:55 crc kubenswrapper[4669]: E1010 09:11:55.795406 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.795410 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:55 crc kubenswrapper[4669]: E1010 09:11:55.795507 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.894238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.894283 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.894294 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.894314 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.894331 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.997095 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.997149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.997160 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.997175 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:55 crc kubenswrapper[4669]: I1010 09:11:55.997186 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:55Z","lastTransitionTime":"2025-10-10T09:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.099337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.099414 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.099448 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.099477 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.099498 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.201718 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.202011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.202021 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.202035 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.202045 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.224277 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/2.log" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.228999 4669 scope.go:117] "RemoveContainer" containerID="ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530" Oct 10 09:11:56 crc kubenswrapper[4669]: E1010 09:11:56.229296 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.257075 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.279826 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.293679 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.306350 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.306407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.306427 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.306450 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.306463 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.309305 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.330169 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.340894 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.352152 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.365082 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.379435 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.395802 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.409921 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.409953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.409961 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.409977 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.409985 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.417182 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.432794 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.447426 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.463232 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.476767 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.486440 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.497000 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:56Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.512913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.512987 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.513013 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.513050 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.513067 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.615793 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.615831 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.615841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.615856 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.615867 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.718673 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.718772 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.718802 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.718826 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.718843 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.794941 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:56 crc kubenswrapper[4669]: E1010 09:11:56.795138 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.821565 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.821662 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.821679 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.821703 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.821719 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.924600 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.924698 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.924715 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.924732 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:56 crc kubenswrapper[4669]: I1010 09:11:56.924743 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:56Z","lastTransitionTime":"2025-10-10T09:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.027627 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.027673 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.027684 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.027701 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.027713 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.129775 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.129835 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.129854 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.129878 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.129894 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.233811 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.233880 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.233903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.233930 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.233945 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.336373 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.336417 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.336428 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.336444 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.336456 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.440334 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.440410 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.440422 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.440466 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.440479 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.543121 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.543154 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.543162 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.543177 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.543186 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.645900 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.645940 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.645949 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.645974 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.645988 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.749276 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.749355 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.749373 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.749403 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.749423 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.794828 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.794897 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.794837 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:57 crc kubenswrapper[4669]: E1010 09:11:57.794969 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:57 crc kubenswrapper[4669]: E1010 09:11:57.795151 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:57 crc kubenswrapper[4669]: E1010 09:11:57.795247 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.811965 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.827414 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.839446 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.850619 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.856861 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.856906 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.856915 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.856932 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.856943 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.862558 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.872570 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.884942 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.901926 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.915720 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.928252 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.941429 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.951923 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.958632 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.958666 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.958678 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.958699 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.958710 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:57Z","lastTransitionTime":"2025-10-10T09:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.961806 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.973262 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.986665 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:57 crc kubenswrapper[4669]: I1010 09:11:57.998114 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:57Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.009778 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:11:58Z is after 2025-08-24T17:21:41Z" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.061749 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.061804 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.061824 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.061855 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.061877 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.164630 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.165576 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.166028 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.166169 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.166295 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.269116 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.269159 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.269171 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.269188 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.269200 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.380546 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.380656 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.380680 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.380710 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.380741 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.483401 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.483438 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.483449 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.483466 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.483478 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.587433 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.587491 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.587505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.587522 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.587535 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.690716 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.691121 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.691306 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.691489 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.691698 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.794342 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.794367 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.794375 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.794407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.794417 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.794811 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:11:58 crc kubenswrapper[4669]: E1010 09:11:58.794935 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.897642 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.897712 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.897727 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.897749 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:58 crc kubenswrapper[4669]: I1010 09:11:58.897781 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:58Z","lastTransitionTime":"2025-10-10T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.000131 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.000205 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.000257 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.000281 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.000991 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.103704 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.103935 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.104053 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.104147 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.104233 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.208123 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.209801 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.209926 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.210031 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.210207 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.313431 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.313760 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.313878 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.314010 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.314104 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.416054 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.416099 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.416110 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.416129 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.416142 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.518891 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.518928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.518937 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.518953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.518963 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.621861 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.621918 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.621938 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.621964 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.621983 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.725704 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.725761 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.725780 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.725803 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.725820 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.795537 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.795539 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.795673 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:11:59 crc kubenswrapper[4669]: E1010 09:11:59.795814 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:11:59 crc kubenswrapper[4669]: E1010 09:11:59.795956 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:11:59 crc kubenswrapper[4669]: E1010 09:11:59.796340 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.829192 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.829236 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.829253 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.829274 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.829293 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.932761 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.932824 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.932841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.932864 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:11:59 crc kubenswrapper[4669]: I1010 09:11:59.932921 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:11:59Z","lastTransitionTime":"2025-10-10T09:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.035964 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.036003 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.036014 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.036032 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.036044 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.137947 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.137985 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.137995 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.138009 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.138020 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.239973 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.240006 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.240038 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.240058 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.240067 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.342895 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.342936 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.342945 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.342959 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.342970 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.445917 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.445976 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.445993 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.446017 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.446051 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.548441 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.548497 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.548511 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.548534 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.548548 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.651431 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.651509 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.651535 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.651563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.651635 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.753863 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.753931 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.753955 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.753988 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.754010 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.794548 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:00 crc kubenswrapper[4669]: E1010 09:12:00.794781 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.857421 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.857486 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.857504 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.857527 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.857544 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.961320 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.961763 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.961872 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.961950 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:00 crc kubenswrapper[4669]: I1010 09:12:00.962027 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:00Z","lastTransitionTime":"2025-10-10T09:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.066093 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.066165 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.066186 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.066214 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.066234 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.169692 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.169735 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.169748 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.169770 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.169783 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.272471 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.272527 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.272540 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.272558 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.272568 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.374626 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.374682 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.374698 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.374718 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.374734 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.477993 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.478040 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.478053 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.478069 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.478081 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.580286 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.580319 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.580330 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.580346 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.580357 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.682711 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.682735 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.682745 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.682757 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.682765 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.785063 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.785089 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.785097 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.785110 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.785118 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.794824 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.794895 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.794848 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:01 crc kubenswrapper[4669]: E1010 09:12:01.794969 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:01 crc kubenswrapper[4669]: E1010 09:12:01.795056 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:01 crc kubenswrapper[4669]: E1010 09:12:01.795171 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.887160 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.887197 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.887206 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.887220 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.887231 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.989200 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.989233 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.989247 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.989263 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:01 crc kubenswrapper[4669]: I1010 09:12:01.989275 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:01Z","lastTransitionTime":"2025-10-10T09:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.092868 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.092901 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.092913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.092928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.092940 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.194703 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.194736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.194746 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.194760 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.194772 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.297258 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.297304 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.297319 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.297337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.297351 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.399771 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.399815 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.399825 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.399842 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.399853 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.502811 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.502862 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.502874 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.502891 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.502902 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.605429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.605456 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.605463 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.605474 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.605482 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.707855 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.707877 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.707885 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.707898 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.707907 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.794698 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:02 crc kubenswrapper[4669]: E1010 09:12:02.794855 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.809755 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.809866 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.809886 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.809907 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.809931 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.912856 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.912928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.912946 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.912973 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:02 crc kubenswrapper[4669]: I1010 09:12:02.912994 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:02Z","lastTransitionTime":"2025-10-10T09:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.015407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.015489 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.015514 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.015543 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.015575 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.117681 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.117718 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.117727 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.117740 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.117749 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.220086 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.220125 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.220135 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.220152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.220165 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.323309 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.323386 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.323486 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.323518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.324143 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.426885 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.426948 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.426966 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.426989 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.427007 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.529816 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.529895 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.529906 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.529928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.529941 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.631946 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.631981 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.631989 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.632003 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.632012 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.735890 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.735929 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.735938 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.735953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.735962 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.796923 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:03 crc kubenswrapper[4669]: E1010 09:12:03.797034 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.797111 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:03 crc kubenswrapper[4669]: E1010 09:12:03.797277 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.797387 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:03 crc kubenswrapper[4669]: E1010 09:12:03.798279 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.838374 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.838402 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.838410 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.838423 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.838432 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.940098 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.940129 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.940138 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.940152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:03 crc kubenswrapper[4669]: I1010 09:12:03.940161 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:03Z","lastTransitionTime":"2025-10-10T09:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.042238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.042273 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.042282 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.042297 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.042309 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.144478 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.144518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.144529 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.144543 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.144554 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.247696 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.247741 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.247754 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.247769 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.247781 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.350232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.350307 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.350318 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.350335 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.350347 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.452701 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.452730 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.452737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.452751 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.452759 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.556170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.556203 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.556215 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.556230 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.556240 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.659007 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.659038 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.659049 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.659065 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.659078 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.761651 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.761682 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.761690 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.761704 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.761712 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.793059 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.793121 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.793139 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.793162 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.793183 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.795371 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:04 crc kubenswrapper[4669]: E1010 09:12:04.795567 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:04 crc kubenswrapper[4669]: E1010 09:12:04.809085 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:04Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.813349 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.813399 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.813409 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.813426 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.813437 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: E1010 09:12:04.827012 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:04Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.830126 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.830157 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.830166 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.830198 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.830212 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: E1010 09:12:04.842744 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:04Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.847412 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.847456 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.847467 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.847483 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.847864 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: E1010 09:12:04.862121 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:04Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.868181 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.868242 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.868256 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.868273 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.868290 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: E1010 09:12:04.881796 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:04Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:04 crc kubenswrapper[4669]: E1010 09:12:04.882036 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.884478 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.884527 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.884539 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.884554 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.884566 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.987491 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.987523 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.987534 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.987549 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:04 crc kubenswrapper[4669]: I1010 09:12:04.987559 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:04Z","lastTransitionTime":"2025-10-10T09:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.091263 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.091332 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.091354 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.091383 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.091404 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.114056 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:05 crc kubenswrapper[4669]: E1010 09:12:05.114214 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:12:05 crc kubenswrapper[4669]: E1010 09:12:05.114328 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:12:37.114304806 +0000 UTC m=+100.130323588 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.194360 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.194393 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.194402 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.194415 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.194424 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.296484 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.296627 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.296647 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.296670 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.296686 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.399518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.399578 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.399631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.399659 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.399681 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.502852 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.502953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.503026 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.503067 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.503084 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.606501 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.606545 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.606557 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.606572 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.606606 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.710764 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.710799 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.710807 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.710821 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.710831 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.794784 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:05 crc kubenswrapper[4669]: E1010 09:12:05.795263 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.794885 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.794883 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:05 crc kubenswrapper[4669]: E1010 09:12:05.795616 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:05 crc kubenswrapper[4669]: E1010 09:12:05.795815 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.813036 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.813267 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.813346 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.813424 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.813493 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.915347 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.915393 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.915407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.915423 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:05 crc kubenswrapper[4669]: I1010 09:12:05.915434 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:05Z","lastTransitionTime":"2025-10-10T09:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.017688 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.017723 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.017733 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.017779 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.017789 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.120620 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.120651 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.120662 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.120675 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.120685 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.223564 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.223621 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.223630 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.223644 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.223655 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.327137 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.327191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.327209 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.327232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.327248 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.429771 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.429821 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.429829 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.429846 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.429871 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.532659 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.532696 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.532727 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.532748 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.532760 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.634939 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.635007 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.635022 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.635046 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.635060 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.737173 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.737204 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.737214 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.737229 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.737241 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.795427 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:06 crc kubenswrapper[4669]: E1010 09:12:06.795688 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.840669 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.840700 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.840712 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.840728 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.840741 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.942620 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.942660 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.942669 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.942683 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:06 crc kubenswrapper[4669]: I1010 09:12:06.942691 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:06Z","lastTransitionTime":"2025-10-10T09:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.045168 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.045225 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.045237 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.045285 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.045301 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.148247 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.148523 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.148738 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.148824 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.148888 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.250419 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.250467 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.250477 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.250491 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.250527 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.267568 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/0.log" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.267670 4669 generic.go:334] "Generic (PLEG): container finished" podID="9502d677-e546-4df7-96c2-bce8274c0f57" containerID="79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620" exitCode=1 Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.267709 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerDied","Data":"79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.268214 4669 scope.go:117] "RemoveContainer" containerID="79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.280644 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.295094 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.306105 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.317687 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.329200 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.339718 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.350571 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.352916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.352961 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.352973 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.352991 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.353003 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.368059 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.384741 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.401577 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.416409 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.428447 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.443228 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.455716 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.455749 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.455759 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.455775 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.455787 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.457560 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.471869 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.485795 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.495147 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.558017 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.558039 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.558047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.558059 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.558067 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.659869 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.659894 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.659902 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.659915 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.659924 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.761369 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.761395 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.761404 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.761416 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.761424 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.794489 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.794626 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:07 crc kubenswrapper[4669]: E1010 09:12:07.794700 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.794709 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:07 crc kubenswrapper[4669]: E1010 09:12:07.795003 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:07 crc kubenswrapper[4669]: E1010 09:12:07.795239 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.805122 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.807032 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.817257 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.833653 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.843153 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.862356 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.864169 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.864209 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.864218 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.864233 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.864243 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.877600 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.896142 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.907270 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.920964 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.937700 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.953043 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.963893 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.966962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.966991 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.967002 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.967017 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.967027 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:07Z","lastTransitionTime":"2025-10-10T09:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.975351 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:07 crc kubenswrapper[4669]: I1010 09:12:07.990622 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.000615 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.008851 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.019800 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.069680 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.069717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.069725 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.069738 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.069747 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.171335 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.171367 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.171377 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.171392 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.171401 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.272846 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/0.log" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.272867 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.272890 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.272900 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.272914 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.272927 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerStarted","Data":"7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.272924 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.285390 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.297320 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.340488 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.360349 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.375150 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.375193 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.375206 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.375223 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.375236 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.383299 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.425998 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.441986 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.455294 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.467072 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.477317 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.477344 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.477353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.477367 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.477375 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.484260 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.497063 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.507063 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.517766 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.531153 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.541225 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63cde3f2-7216-4708-8553-72bfbae870f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.553512 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.569237 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.579417 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.579455 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.579466 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.579480 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.579489 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.581565 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.681278 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.681304 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.681311 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.681323 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.681331 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.783073 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.783257 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.783360 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.783452 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.783532 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.794261 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:08 crc kubenswrapper[4669]: E1010 09:12:08.794346 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.794924 4669 scope.go:117] "RemoveContainer" containerID="ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530" Oct 10 09:12:08 crc kubenswrapper[4669]: E1010 09:12:08.795034 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.885149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.885176 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.885184 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.885195 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.885203 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.987563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.987699 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.987722 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.987751 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:08 crc kubenswrapper[4669]: I1010 09:12:08.987771 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:08Z","lastTransitionTime":"2025-10-10T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.090183 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.090407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.090468 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.090529 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.090619 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.192733 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.192820 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.192839 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.192861 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.192878 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.294757 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.294784 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.294792 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.294805 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.294814 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.396506 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.396551 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.396563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.396612 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.396624 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.498059 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.498099 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.498111 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.498125 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.498134 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.600314 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.600346 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.600354 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.600368 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.600376 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.702686 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.702729 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.702741 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.702756 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.702766 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.794692 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.794736 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.794729 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:09 crc kubenswrapper[4669]: E1010 09:12:09.794834 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:09 crc kubenswrapper[4669]: E1010 09:12:09.795039 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:09 crc kubenswrapper[4669]: E1010 09:12:09.795126 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.804717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.804778 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.804802 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.804831 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.804854 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.907504 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.907564 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.907581 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.907628 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:09 crc kubenswrapper[4669]: I1010 09:12:09.907644 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:09Z","lastTransitionTime":"2025-10-10T09:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.010166 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.010195 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.010204 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.010218 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.010228 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.113104 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.113166 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.113180 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.113206 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.113222 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.216378 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.216426 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.216438 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.216460 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.216475 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.318631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.318665 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.318674 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.318687 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.318697 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.421220 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.421264 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.421277 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.421295 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.421310 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.524022 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.524064 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.524072 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.524086 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.524096 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.626723 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.626768 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.626783 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.626799 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.626812 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.729342 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.729384 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.729394 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.729413 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.729422 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.797133 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:10 crc kubenswrapper[4669]: E1010 09:12:10.797331 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.831124 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.831454 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.831672 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.831708 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.831726 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.934810 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.934837 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.934845 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.934859 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:10 crc kubenswrapper[4669]: I1010 09:12:10.934868 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:10Z","lastTransitionTime":"2025-10-10T09:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.037842 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.038105 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.038206 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.038294 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.038386 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.140844 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.140896 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.140915 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.140940 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.140956 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.243474 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.243547 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.243568 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.243631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.243654 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.346391 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.346421 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.346432 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.346446 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.346457 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.449211 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.449249 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.449262 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.449278 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.449289 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.552000 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.553090 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.553344 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.553547 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.553727 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.656821 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.657124 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.657207 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.657304 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.657402 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.760574 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.760662 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.760711 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.760736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.760755 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.794681 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.794742 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.794686 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:11 crc kubenswrapper[4669]: E1010 09:12:11.794918 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:11 crc kubenswrapper[4669]: E1010 09:12:11.795001 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:11 crc kubenswrapper[4669]: E1010 09:12:11.795154 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.863018 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.863064 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.863076 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.863092 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.863102 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.965426 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.965702 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.965782 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.965849 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:11 crc kubenswrapper[4669]: I1010 09:12:11.965910 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:11Z","lastTransitionTime":"2025-10-10T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.077477 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.077581 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.077662 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.077697 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.077721 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.181045 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.181152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.181165 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.181190 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.181202 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.285344 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.285386 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.285402 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.285424 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.285440 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.389194 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.389252 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.389265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.389288 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.389304 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.492753 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.492846 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.492863 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.492886 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.492902 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.595730 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.595855 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.595882 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.595928 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.595952 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.698349 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.698415 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.698436 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.698459 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.698475 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.794768 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:12 crc kubenswrapper[4669]: E1010 09:12:12.794958 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.800901 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.800953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.800970 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.800990 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.801006 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.903000 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.903036 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.903047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.903062 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:12 crc kubenswrapper[4669]: I1010 09:12:12.903072 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:12Z","lastTransitionTime":"2025-10-10T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.006016 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.006086 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.006108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.006132 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.006144 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.109556 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.109610 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.109621 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.109669 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.109683 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.212763 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.212818 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.212830 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.212851 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.212866 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.325087 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.325137 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.325152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.325176 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.325192 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.429575 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.429741 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.429765 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.429794 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.429816 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.532223 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.532264 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.532276 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.532292 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.532303 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.635397 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.635505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.635530 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.635557 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.635578 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.738493 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.738559 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.738580 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.738639 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.738659 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.795236 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:13 crc kubenswrapper[4669]: E1010 09:12:13.795459 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.795503 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.795255 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:13 crc kubenswrapper[4669]: E1010 09:12:13.795715 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:13 crc kubenswrapper[4669]: E1010 09:12:13.795895 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.841559 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.841650 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.841668 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.841691 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.841709 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.945616 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.945674 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.945691 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.945716 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:13 crc kubenswrapper[4669]: I1010 09:12:13.945736 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:13Z","lastTransitionTime":"2025-10-10T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.049094 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.049199 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.049221 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.049246 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.049264 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.152427 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.152635 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.152687 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.152737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.152753 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.257297 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.257356 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.257374 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.257396 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.257414 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.361353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.361454 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.361506 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.361532 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.361550 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.465485 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.465528 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.465549 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.465566 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.465578 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.571025 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.571082 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.571105 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.571135 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.571155 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.674359 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.674421 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.674433 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.674453 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.674473 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.778005 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.778070 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.778108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.778140 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.778162 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.795405 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:14 crc kubenswrapper[4669]: E1010 09:12:14.795611 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.881041 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.881115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.881150 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.881182 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.881203 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.984940 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.985003 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.985019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.985043 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:14 crc kubenswrapper[4669]: I1010 09:12:14.985063 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:14Z","lastTransitionTime":"2025-10-10T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.088538 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.088646 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.088666 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.088693 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.088711 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.170712 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.170783 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.170802 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.170829 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.170849 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.186334 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.191129 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.191190 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.191214 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.191239 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.191256 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.215616 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.221116 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.221176 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.221195 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.221219 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.221236 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.240043 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.245308 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.245447 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.245546 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.245673 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.245771 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.265986 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.270320 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.270359 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.270394 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.270412 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.270423 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.285355 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.285548 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.288200 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.288252 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.288264 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.288279 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.288291 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.390165 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.390206 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.390221 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.390240 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.390253 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.493418 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.493472 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.493486 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.493505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.493518 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.595697 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.595741 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.595752 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.595770 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.595779 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.699333 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.699378 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.699387 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.699404 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.699417 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.794486 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.794724 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.794786 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.794913 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.795139 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:15 crc kubenswrapper[4669]: E1010 09:12:15.795272 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.801397 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.801452 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.801464 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.801481 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.801491 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.904440 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.904494 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.904521 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.904543 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:15 crc kubenswrapper[4669]: I1010 09:12:15.904554 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:15Z","lastTransitionTime":"2025-10-10T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.007460 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.007511 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.007522 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.007542 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.007552 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.111337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.111391 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.111403 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.111424 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.111437 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.214397 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.214459 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.214476 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.214503 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.214525 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.317791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.317853 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.317869 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.317897 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.317916 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.420869 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.421029 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.421119 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.421151 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.421175 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.524319 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.524413 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.524430 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.524447 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.524459 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.627434 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.627489 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.627511 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.627532 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.627547 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.731182 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.731264 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.731289 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.731324 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.731352 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.794818 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:16 crc kubenswrapper[4669]: E1010 09:12:16.795066 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.834308 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.834362 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.834380 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.834407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.834430 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.937806 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.937864 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.937881 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.937906 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:16 crc kubenswrapper[4669]: I1010 09:12:16.937962 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:16Z","lastTransitionTime":"2025-10-10T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.041163 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.041239 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.041264 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.041291 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.041321 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.149505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.149621 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.149641 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.149664 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.149685 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.252331 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.252401 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.252425 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.252454 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.252475 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.355693 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.355736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.355752 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.355774 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.355790 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.457791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.457839 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.457856 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.457879 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.457894 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.560631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.560672 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.560684 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.560699 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.560710 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.663173 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.663227 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.663237 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.663256 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.663267 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.765659 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.765697 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.765707 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.765725 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.765735 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.795344 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.795430 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.795453 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:17 crc kubenswrapper[4669]: E1010 09:12:17.795674 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:17 crc kubenswrapper[4669]: E1010 09:12:17.795976 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:17 crc kubenswrapper[4669]: E1010 09:12:17.796048 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.814173 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63cde3f2-7216-4708-8553-72bfbae870f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.834965 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.856145 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.868692 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.868736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.868762 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.868786 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.868810 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.869847 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.883985 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.899100 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.913041 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.927174 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.950769 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.966691 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.972489 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.972524 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.972532 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.972546 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.972554 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:17Z","lastTransitionTime":"2025-10-10T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.981522 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:17 crc kubenswrapper[4669]: I1010 09:12:17.996108 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.012763 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.028682 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.048397 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.066031 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.075518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.075647 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.075679 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.075708 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.075733 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.082928 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.094444 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.178511 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.178551 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.178562 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.178631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.178644 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.285990 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.286047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.286066 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.286090 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.286108 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.389053 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.389134 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.389153 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.389178 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.389195 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.493101 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.493181 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.493196 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.493240 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.493278 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.596527 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.596616 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.596634 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.596660 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.596681 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.699755 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.699813 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.699830 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.699852 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.699868 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.794811 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:18 crc kubenswrapper[4669]: E1010 09:12:18.795022 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.802554 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.802601 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.802612 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.802631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.802642 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.905343 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.905414 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.905431 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.905457 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:18 crc kubenswrapper[4669]: I1010 09:12:18.905470 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:18Z","lastTransitionTime":"2025-10-10T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.008740 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.008787 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.008804 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.008827 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.008843 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.111405 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.111472 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.111495 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.111525 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.111548 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.214493 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.214563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.214642 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.214680 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.214716 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.317316 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.317376 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.317400 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.317427 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.317448 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.420417 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.420481 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.420504 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.420534 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.420556 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.523202 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.523274 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.523312 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.523353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.523365 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.626679 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.626714 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.626725 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.626739 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.626750 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.729676 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.729708 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.729717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.729730 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.729739 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.794570 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.794742 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:19 crc kubenswrapper[4669]: E1010 09:12:19.794872 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.795140 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:19 crc kubenswrapper[4669]: E1010 09:12:19.795186 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:19 crc kubenswrapper[4669]: E1010 09:12:19.795187 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.815292 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.832746 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.832822 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.832843 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.832870 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.832888 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.935865 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.935895 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.935905 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.935920 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:19 crc kubenswrapper[4669]: I1010 09:12:19.935932 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:19Z","lastTransitionTime":"2025-10-10T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.039229 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.039322 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.039341 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.039365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.039383 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.143111 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.143170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.143188 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.143217 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.143235 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.245931 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.246075 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.246162 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.246193 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.246211 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.349015 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.349073 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.349091 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.349149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.349167 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.452210 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.452252 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.452262 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.452284 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.452296 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.554910 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.554945 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.554954 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.554968 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.554977 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.658641 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.658693 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.658711 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.658734 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.658751 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.762183 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.762239 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.762250 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.762269 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.762281 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.795365 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:20 crc kubenswrapper[4669]: E1010 09:12:20.795859 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.865814 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.865871 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.865889 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.865911 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.865929 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.969643 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.969699 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.969716 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.969739 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:20 crc kubenswrapper[4669]: I1010 09:12:20.969757 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:20Z","lastTransitionTime":"2025-10-10T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.072519 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.072641 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.072666 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.072696 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.072720 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.176369 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.176473 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.176499 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.176529 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.176552 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.280041 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.280102 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.280122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.280148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.280166 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.383277 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.383336 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.383351 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.383370 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.383383 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.485793 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.485833 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.485846 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.485864 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.485878 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.587917 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.588022 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.588033 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.588048 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.588058 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.597686 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.597832 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.597811846 +0000 UTC m=+148.613830598 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.597896 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.597933 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.598116 4669 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.598175 4669 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.598207 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.598184607 +0000 UTC m=+148.614203389 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.598235 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.598222818 +0000 UTC m=+148.614241600 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.690438 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.690521 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.690539 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.690560 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.690576 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.699321 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.699386 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699484 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699501 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699511 4669 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699553 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.699540911 +0000 UTC m=+148.715559653 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699563 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699611 4669 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699630 4669 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.699681 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.699662384 +0000 UTC m=+148.715681156 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.793337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.793398 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.793416 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.793441 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.793459 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.794673 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.794763 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.794912 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.795554 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.795668 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:21 crc kubenswrapper[4669]: E1010 09:12:21.795745 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.796333 4669 scope.go:117] "RemoveContainer" containerID="ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.896150 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.896190 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.896224 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.896243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.896255 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.998855 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.998882 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.998893 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.998908 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:21 crc kubenswrapper[4669]: I1010 09:12:21.998919 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:21Z","lastTransitionTime":"2025-10-10T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.101491 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.101519 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.101527 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.101541 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.101550 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.204365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.204397 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.204408 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.204425 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.204436 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.306393 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.306429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.306440 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.306452 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.306460 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.321702 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/2.log" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.323754 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.324770 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.340288 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.366278 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.383484 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.395858 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.405063 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.409550 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.409610 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.409622 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.409639 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.409651 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.423387 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.437294 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.447737 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.458454 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.468242 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.482043 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.498064 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73b6e8ae-172b-414b-b3ff-ce74edf7954b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1db2b1fc6bf595859b800a987b85701bd3e814c627c1795a573ac49b8fd621e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08061d946fd75010d0a38811215a27587ead13f95b1fd145a4011e19d8f96f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://766207b5eb0a9a74e33f9401e4e7973100af2d4fa385cef24239061db0e81ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f5cb9f2e8b5efefffb5a810fc2bf904ef0ae855b723207b4d0b2b024ce6571\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f30092148564b7d90704c6cb8df802f66670871f12b3fac0ade6a238144ea204\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.509124 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.511903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.511931 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.511939 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.511953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.511964 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.518382 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.526384 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.538261 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.547476 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63cde3f2-7216-4708-8553-72bfbae870f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.565579 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.580897 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:22Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.614061 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.614091 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.614099 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.614111 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.614120 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.716208 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.716236 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.716244 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.716257 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.716264 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.794971 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:22 crc kubenswrapper[4669]: E1010 09:12:22.795445 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.818011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.818041 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.818051 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.818065 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.818082 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.920896 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.921167 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.921284 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.921358 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:22 crc kubenswrapper[4669]: I1010 09:12:22.921452 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:22Z","lastTransitionTime":"2025-10-10T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.024312 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.024360 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.024377 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.024396 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.024412 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.127618 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.127675 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.127694 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.127719 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.127736 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.230256 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.230356 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.230381 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.230416 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.230435 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.328852 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/3.log" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.329626 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/2.log" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.332319 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.332365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.332377 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.332394 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.332408 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.333947 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" exitCode=1 Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.334014 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.334089 4669 scope.go:117] "RemoveContainer" containerID="ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.336381 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:12:23 crc kubenswrapper[4669]: E1010 09:12:23.336905 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.350486 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.364651 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.376713 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.388918 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.400699 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.412798 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.426338 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.434763 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.434910 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.434971 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.435067 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.435154 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.446855 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73b6e8ae-172b-414b-b3ff-ce74edf7954b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1db2b1fc6bf595859b800a987b85701bd3e814c627c1795a573ac49b8fd621e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08061d946fd75010d0a38811215a27587ead13f95b1fd145a4011e19d8f96f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://766207b5eb0a9a74e33f9401e4e7973100af2d4fa385cef24239061db0e81ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f5cb9f2e8b5efefffb5a810fc2bf904ef0ae855b723207b4d0b2b024ce6571\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f30092148564b7d90704c6cb8df802f66670871f12b3fac0ade6a238144ea204\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.460848 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.470936 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.479868 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.489908 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.499563 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63cde3f2-7216-4708-8553-72bfbae870f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.511777 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.522892 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.531657 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.536851 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.536889 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.536899 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.536916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.536928 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.540722 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.551313 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.569661 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec0b653ebdec79a64f1907ff875dd837cf08866dd99dc3471d7e47487743a530\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"message\\\":\\\"ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.219\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1010 09:11:54.722577 6313 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1010 09:11:54.722555 6313 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:22Z\\\",\\\"message\\\":\\\"ster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561279 6679 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561286 6679 services_controller.go:453] Built service openshift-kube-scheduler-operator/metrics template LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561293 6679 services_controller.go:454] Service openshift-kube-scheduler-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561308 6679 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-bhn82\\\\nI1010 09:12:22.561304 6679 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"1dc899db-4498-4b7a-8437-861940b962e7\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:12:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.639739 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.639795 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.639807 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.639822 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.639833 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.743235 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.743303 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.743326 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.743353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.743374 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.795068 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.795152 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:23 crc kubenswrapper[4669]: E1010 09:12:23.795278 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.795066 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:23 crc kubenswrapper[4669]: E1010 09:12:23.795433 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:23 crc kubenswrapper[4669]: E1010 09:12:23.795528 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.847519 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.848481 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.848736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.848903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.849038 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.951386 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.951429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.951441 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.951458 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:23 crc kubenswrapper[4669]: I1010 09:12:23.951470 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:23Z","lastTransitionTime":"2025-10-10T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.054541 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.054655 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.054674 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.054733 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.054751 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.157646 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.157720 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.157740 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.157814 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.157833 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.261328 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.261460 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.261485 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.261518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.261539 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.340395 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/3.log" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.344560 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:12:24 crc kubenswrapper[4669]: E1010 09:12:24.344719 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.364303 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.364362 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.364379 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.364404 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.364422 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.380805 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73b6e8ae-172b-414b-b3ff-ce74edf7954b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1db2b1fc6bf595859b800a987b85701bd3e814c627c1795a573ac49b8fd621e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08061d946fd75010d0a38811215a27587ead13f95b1fd145a4011e19d8f96f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://766207b5eb0a9a74e33f9401e4e7973100af2d4fa385cef24239061db0e81ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f5cb9f2e8b5efefffb5a810fc2bf904ef0ae855b723207b4d0b2b024ce6571\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f30092148564b7d90704c6cb8df802f66670871f12b3fac0ade6a238144ea204\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.399048 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.409853 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.425515 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.435927 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.447558 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.456328 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.464548 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.471526 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.471802 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.471924 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.472044 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.472152 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.474780 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.483092 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63cde3f2-7216-4708-8553-72bfbae870f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.493129 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.508425 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:22Z\\\",\\\"message\\\":\\\"ster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561279 6679 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561286 6679 services_controller.go:453] Built service openshift-kube-scheduler-operator/metrics template LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561293 6679 services_controller.go:454] Service openshift-kube-scheduler-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561308 6679 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-bhn82\\\\nI1010 09:12:22.561304 6679 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"1dc899db-4498-4b7a-8437-861940b962e7\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:12:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.518406 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.528246 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.538497 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.548780 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.557542 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.571687 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.573895 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.573917 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.573926 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.573941 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.573953 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.583535 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:24Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.677147 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.677469 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.677656 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.677857 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.678088 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.781053 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.781125 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.781163 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.781198 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.781233 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.795533 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:24 crc kubenswrapper[4669]: E1010 09:12:24.795866 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.884746 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.884789 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.884801 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.884818 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.884830 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.988649 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.988699 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.988710 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.988730 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:24 crc kubenswrapper[4669]: I1010 09:12:24.988742 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:24Z","lastTransitionTime":"2025-10-10T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.092169 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.092265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.092297 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.092325 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.092347 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.196042 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.196114 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.196129 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.196152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.196170 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.299242 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.299955 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.299986 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.300013 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.300050 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.403488 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.403533 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.403543 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.403558 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.403571 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.505204 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.505240 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.505250 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.505266 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.505277 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.608049 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.608081 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.608092 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.608108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.608119 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.636293 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.636339 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.636350 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.636366 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.636379 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.651118 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.654651 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.654701 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.654717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.654737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.654751 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.669375 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.674130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.674184 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.674203 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.674226 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.674244 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.693120 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.696496 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.696546 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.696561 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.696586 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.696619 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.711181 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.714846 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.714878 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.714889 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.714906 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.714918 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.727970 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.728121 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.729593 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.729652 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.729669 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.729690 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.729705 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.794788 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.795193 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.794847 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.795541 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.794788 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:25 crc kubenswrapper[4669]: E1010 09:12:25.795851 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.832122 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.832153 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.832161 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.832172 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.832181 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.934224 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.934276 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.934292 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.934318 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:25 crc kubenswrapper[4669]: I1010 09:12:25.934332 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:25Z","lastTransitionTime":"2025-10-10T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.036631 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.037002 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.037158 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.037286 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.037400 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.140961 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.141023 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.141048 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.141078 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.141103 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.244284 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.244344 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.244365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.244392 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.244412 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.346468 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.346512 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.346526 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.346542 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.346553 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.449983 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.450272 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.450379 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.450510 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.450610 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.553667 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.553999 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.554094 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.554196 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.554293 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.656638 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.656672 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.656680 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.656696 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.656705 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.759174 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.759206 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.759214 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.759230 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.759238 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.795436 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:26 crc kubenswrapper[4669]: E1010 09:12:26.795663 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.862426 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.862500 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.862550 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.862583 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.862674 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.965475 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.965520 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.965532 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.965549 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:26 crc kubenswrapper[4669]: I1010 09:12:26.965563 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:26Z","lastTransitionTime":"2025-10-10T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.068463 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.068512 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.068525 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.068552 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.068568 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.171705 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.171745 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.171759 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.171778 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.171794 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.275070 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.275295 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.275462 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.275600 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.275746 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.378579 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.378962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.379164 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.379317 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.379441 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.482149 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.482500 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.482516 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.482536 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.482552 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.586340 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.586403 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.586419 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.586441 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.586458 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.689147 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.689202 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.689216 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.689232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.689244 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.791041 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.791324 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.791402 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.791513 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.791620 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.794459 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.794581 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.794577 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:27 crc kubenswrapper[4669]: E1010 09:12:27.795231 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:27 crc kubenswrapper[4669]: E1010 09:12:27.795317 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:27 crc kubenswrapper[4669]: E1010 09:12:27.795376 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.811993 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.824022 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.840810 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.855389 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.866278 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63cde3f2-7216-4708-8553-72bfbae870f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.883344 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.894032 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.894058 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.894082 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.894095 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.894106 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.904302 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:22Z\\\",\\\"message\\\":\\\"ster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561279 6679 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561286 6679 services_controller.go:453] Built service openshift-kube-scheduler-operator/metrics template LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561293 6679 services_controller.go:454] Service openshift-kube-scheduler-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561308 6679 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-bhn82\\\\nI1010 09:12:22.561304 6679 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"1dc899db-4498-4b7a-8437-861940b962e7\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:12:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.915883 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.927197 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.937914 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.947732 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.956707 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.971720 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.982850 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.992112 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:27Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.995767 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.995811 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.995823 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.995840 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:27 crc kubenswrapper[4669]: I1010 09:12:27.995851 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:27Z","lastTransitionTime":"2025-10-10T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.005248 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:28Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.018549 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:28Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.035070 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:28Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.059281 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73b6e8ae-172b-414b-b3ff-ce74edf7954b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1db2b1fc6bf595859b800a987b85701bd3e814c627c1795a573ac49b8fd621e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08061d946fd75010d0a38811215a27587ead13f95b1fd145a4011e19d8f96f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://766207b5eb0a9a74e33f9401e4e7973100af2d4fa385cef24239061db0e81ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f5cb9f2e8b5efefffb5a810fc2bf904ef0ae855b723207b4d0b2b024ce6571\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f30092148564b7d90704c6cb8df802f66670871f12b3fac0ade6a238144ea204\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:28Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.098082 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.098121 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.098130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.098145 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.098154 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.201071 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.201140 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.201160 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.201184 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.201201 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.311226 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.311287 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.311305 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.311328 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.311347 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.413743 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.413804 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.413826 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.413851 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.413870 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.516911 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.516980 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.516997 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.517023 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.517041 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.620765 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.620826 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.620851 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.620880 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.620903 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.724224 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.724280 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.724295 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.724323 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.724363 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.795240 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:28 crc kubenswrapper[4669]: E1010 09:12:28.795440 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.828120 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.828185 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.828209 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.828238 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.828260 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.930944 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.931002 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.931013 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.931042 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:28 crc kubenswrapper[4669]: I1010 09:12:28.931053 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:28Z","lastTransitionTime":"2025-10-10T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.034316 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.034401 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.034425 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.034459 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.034482 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.136980 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.137030 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.137044 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.137065 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.137081 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.239342 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.239381 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.239392 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.239408 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.239419 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.342901 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.342941 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.342957 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.342980 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.342994 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.445453 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.445508 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.445518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.445535 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.445546 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.548901 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.548966 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.548981 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.548999 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.549015 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.652142 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.652306 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.652400 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.652476 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.652502 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.756442 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.757204 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.757454 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.757788 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.758002 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.794982 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:29 crc kubenswrapper[4669]: E1010 09:12:29.795411 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.795114 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:29 crc kubenswrapper[4669]: E1010 09:12:29.795975 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.795044 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:29 crc kubenswrapper[4669]: E1010 09:12:29.796423 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.860942 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.860990 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.861002 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.861021 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.861032 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.963373 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.963412 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.963421 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.963434 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:29 crc kubenswrapper[4669]: I1010 09:12:29.963444 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:29Z","lastTransitionTime":"2025-10-10T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.066894 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.066957 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.066973 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.066996 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.067013 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.169892 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.169946 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.169962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.169984 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.170001 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.273200 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.273262 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.273282 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.273309 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.273329 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.375913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.375964 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.375981 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.376003 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.376020 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.479266 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.479657 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.479836 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.479977 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.480107 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.583047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.583108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.583131 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.583156 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.583187 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.686143 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.686215 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.686236 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.686265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.686286 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.789746 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.789818 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.789840 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.789872 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.789892 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.794405 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:30 crc kubenswrapper[4669]: E1010 09:12:30.794654 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.893909 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.893962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.893983 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.894011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.894034 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.997520 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.997626 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.997650 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.997674 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:30 crc kubenswrapper[4669]: I1010 09:12:30.997692 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:30Z","lastTransitionTime":"2025-10-10T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.100678 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.100733 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.100750 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.100773 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.100790 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.203888 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.203953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.203975 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.204006 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.204042 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.307141 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.307191 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.307207 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.307227 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.307240 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.410869 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.410940 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.410960 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.410985 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.411002 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.513915 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.513980 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.513998 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.514425 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.514478 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.617351 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.617385 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.617396 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.617412 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.617424 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.720397 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.720465 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.720480 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.720503 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.720537 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.794880 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.794956 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:31 crc kubenswrapper[4669]: E1010 09:12:31.795074 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.794956 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:31 crc kubenswrapper[4669]: E1010 09:12:31.795233 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:31 crc kubenswrapper[4669]: E1010 09:12:31.795352 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.823790 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.823832 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.823850 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.823872 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.823890 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.925880 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.925951 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.925965 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.925985 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:31 crc kubenswrapper[4669]: I1010 09:12:31.925999 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:31Z","lastTransitionTime":"2025-10-10T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.029218 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.029255 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.029267 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.029284 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.029294 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.131204 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.131244 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.131253 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.131266 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.131276 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.234047 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.234125 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.234145 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.234170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.234189 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.337204 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.337282 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.337309 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.337347 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.337376 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.440471 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.440546 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.440567 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.440623 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.440645 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.544909 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.544978 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.545000 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.545029 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.545048 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.651784 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.651826 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.651835 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.651853 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.651865 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.754368 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.754413 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.754436 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.754456 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.754471 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.794490 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:32 crc kubenswrapper[4669]: E1010 09:12:32.794737 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.856916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.856986 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.857011 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.857039 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.857060 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.959744 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.959779 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.959791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.959808 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:32 crc kubenswrapper[4669]: I1010 09:12:32.959820 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:32Z","lastTransitionTime":"2025-10-10T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.063292 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.063349 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.063365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.063389 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.063408 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.166294 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.166339 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.166349 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.166389 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.166399 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.270088 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.270158 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.270170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.270190 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.270203 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.373485 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.373527 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.373543 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.373561 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.373575 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.476819 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.476859 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.476871 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.476890 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.476902 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.579988 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.580152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.580172 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.580227 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.580251 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.683656 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.683820 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.683838 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.683860 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.683903 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.787143 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.787214 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.787236 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.787264 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.787285 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.794811 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.794888 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.794825 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:33 crc kubenswrapper[4669]: E1010 09:12:33.795010 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:33 crc kubenswrapper[4669]: E1010 09:12:33.795217 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:33 crc kubenswrapper[4669]: E1010 09:12:33.795364 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.889572 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.889634 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.889645 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.889662 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.889675 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.992361 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.992401 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.992412 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.992428 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:33 crc kubenswrapper[4669]: I1010 09:12:33.992439 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:33Z","lastTransitionTime":"2025-10-10T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.095447 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.095485 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.095493 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.095508 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.095518 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.197800 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.197839 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.197849 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.197866 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.197880 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.300688 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.300733 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.300751 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.300773 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.300790 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.403562 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.403664 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.403690 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.403713 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.403730 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.506136 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.506193 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.506215 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.506243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.506265 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.608755 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.608811 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.608823 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.608906 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.608948 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.711463 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.711521 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.711533 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.711548 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.711559 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.795391 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:34 crc kubenswrapper[4669]: E1010 09:12:34.795568 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.813353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.813404 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.813416 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.813436 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.813448 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.915429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.915688 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.915794 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.915871 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:34 crc kubenswrapper[4669]: I1010 09:12:34.915933 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:34Z","lastTransitionTime":"2025-10-10T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.018976 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.019233 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.019335 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.019444 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.019561 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.122036 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.122088 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.122108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.122132 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.122150 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.225523 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.226437 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.226557 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.226699 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.226832 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.329841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.329924 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.329940 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.329958 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.329971 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.432925 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.433016 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.433025 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.433042 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.433057 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.535775 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.536083 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.536220 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.536369 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.536706 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.639488 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.639552 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.639575 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.639644 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.639671 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.741888 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.742242 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.742569 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.742918 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.743038 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.765753 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.770999 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.771101 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.771126 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.771159 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.771194 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.793115 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.795183 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.795914 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.796146 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.796241 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.796468 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.796640 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.799834 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.799886 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.799909 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.799937 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.799958 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.819002 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.824520 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.824575 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.824671 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.824700 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.824724 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.844363 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.850468 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.850775 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.850879 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.851024 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.851137 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.871372 4669 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"dc51e5f1-4956-4174-82c2-97dec3a7897f\\\",\\\"systemUUID\\\":\\\"fd4c383f-0ffa-443a-a38f-89337f2bc2e5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:35 crc kubenswrapper[4669]: E1010 09:12:35.871489 4669 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.873126 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.873193 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.873213 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.873239 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.873257 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.975438 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.975498 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.975516 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.975541 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:35 crc kubenswrapper[4669]: I1010 09:12:35.975559 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:35Z","lastTransitionTime":"2025-10-10T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.078987 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.079027 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.079039 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.079078 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.079088 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.182042 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.182115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.182138 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.182168 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.182189 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.284914 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.284951 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.284962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.284978 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.284989 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.387685 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.387723 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.387734 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.387749 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.387760 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.490312 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.490353 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.490364 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.490381 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.490394 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.617105 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.617142 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.617154 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.617170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.617182 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.719781 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.719813 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.719824 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.719841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.719852 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.795078 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:12:36 crc kubenswrapper[4669]: E1010 09:12:36.795212 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.795370 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:36 crc kubenswrapper[4669]: E1010 09:12:36.795424 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.823705 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.823766 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.823784 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.823806 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.823822 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.927130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.927189 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.927208 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.927232 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:36 crc kubenswrapper[4669]: I1010 09:12:36.927249 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:36Z","lastTransitionTime":"2025-10-10T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.031082 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.031167 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.031188 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.031222 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.031246 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.134061 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.134128 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.134146 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.134169 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.134186 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.162162 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:37 crc kubenswrapper[4669]: E1010 09:12:37.162332 4669 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:12:37 crc kubenswrapper[4669]: E1010 09:12:37.162436 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs podName:20d5735d-7ca2-4824-9b5d-4bb39502a3dc nodeName:}" failed. No retries permitted until 2025-10-10 09:13:41.162404 +0000 UTC m=+164.178422782 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs") pod "network-metrics-daemon-bhn82" (UID: "20d5735d-7ca2-4824-9b5d-4bb39502a3dc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.237077 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.237136 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.237147 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.237170 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.237182 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.340463 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.340516 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.340529 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.340552 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.340566 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.443147 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.443240 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.443255 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.443291 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.443310 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.546320 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.546366 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.546383 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.546407 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.546424 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.649835 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.649899 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.649916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.649940 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.649958 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.752471 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.752505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.752514 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.752528 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.752537 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.795229 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:37 crc kubenswrapper[4669]: E1010 09:12:37.795332 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.795463 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:37 crc kubenswrapper[4669]: E1010 09:12:37.795516 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.795646 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:37 crc kubenswrapper[4669]: E1010 09:12:37.795812 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.817659 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0176566e-cd82-437a-a318-c8046b69903c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea723460b61fdb427a891b381d96292d5bec87a367da29d2b7494a8b87707ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://216a30e8bf5880a54e7e63adf8a52d72115be7a18f6ef0ff8171110a349a82ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f8235d9eed0b3926ebe38270d8d3b8c5b80a9d2948c841ab1f1bce6b6a59b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2465acaaec342404a3c03ef4d03f9f38e30b3a1012296583633fc6561db8c99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91cbaa75e65f468f1d27336bc0a2b46f76dc814d656e9b3ff84645e0b1b18b02\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"g file observer\\\\nW1010 09:11:16.974347 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1010 09:11:16.974504 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 09:11:16.977345 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-32381004/tls.crt::/tmp/serving-cert-32381004/tls.key\\\\\\\"\\\\nI1010 09:11:17.779346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 09:11:17.793207 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 09:11:17.802844 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 09:11:17.802988 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 09:11:17.803859 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 09:11:17.811419 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1010 09:11:17.823820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 09:11:17.823981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 09:11:17.824007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 09:11:17.824067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 09:11:17.824094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1010 09:11:17.813702 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1010 09:11:17.813846 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d658ab506d4ef94e715d972431943b7dfca6b0203f9a4e051c35d4184ec365c2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55efcf74e11dfa227826b67e059b9f4b004dfeb26005c14924213e588ce5f9fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.845464 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73b6e8ae-172b-414b-b3ff-ce74edf7954b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1db2b1fc6bf595859b800a987b85701bd3e814c627c1795a573ac49b8fd621e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08061d946fd75010d0a38811215a27587ead13f95b1fd145a4011e19d8f96f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://766207b5eb0a9a74e33f9401e4e7973100af2d4fa385cef24239061db0e81ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6f5cb9f2e8b5efefffb5a810fc2bf904ef0ae855b723207b4d0b2b024ce6571\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f30092148564b7d90704c6cb8df802f66670871f12b3fac0ade6a238144ea204\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4710373ea5944bdea977dc2cbb97f0314cc5104816ea77a09f2a98264c906aef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff0336c008c7cb72811a2799a5f8f9b3cded4d06ea40b87a61a4befea287cca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9699e719d632022eb54f1ec8ba269afb23527cae573c44ad2aebc58e5b1ee8ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.855354 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.855381 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.855389 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.855403 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.855412 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.860199 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nq6jn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9502d677-e546-4df7-96c2-bce8274c0f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:06Z\\\",\\\"message\\\":\\\"2025-10-10T09:11:21+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c\\\\n2025-10-10T09:11:21+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_182891e7-ec07-4593-bcda-f54a4512ea5c to /host/opt/cni/bin/\\\\n2025-10-10T09:11:21Z [verbose] multus-daemon started\\\\n2025-10-10T09:11:21Z [verbose] Readiness Indicator file check\\\\n2025-10-10T09:12:06Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fm2ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nq6jn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.873275 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-bhn82" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nc9bg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:33Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-bhn82\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.890715 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pvhp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb764f3-806c-4614-b6a4-f247ff3fa796\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16f5af6cea57d071331afd3085411a05a2db8906819eb70bbea21fd2a87de31f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tktp2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pvhp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.902011 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eaa7fbe4-1c87-4c6b-befc-64a1473973b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74bdbfd818497cf8cdbb3e3310a9623960bbdc49283dc67aadf04ac5470c7d5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2feaa933990faad575d06391931c43a4eaec6cdabe29a33448726cc25eda3b3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2d4st\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2v44p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.911624 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63cde3f2-7216-4708-8553-72bfbae870f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67cedc0d8991b57990e1b72156a088e938886d985996b7af498314cb9287091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b309efd89737b15f005a8301546880355d18663b0f7844b50235c276f52ac964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.927968 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7325bc6ee0ce5b591b3129443668399ed20caef5542ad30542eb3f4e3e7d995b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.943946 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6583f591-656e-4230-9772-b2cd0d5176c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e43c359e64d2dc96c67acdbacd714958789f7539aaefe1338da3cb7e66e27111\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef2b560b43bc0e931baabe46e7eb709a85b164083359fe208f60cc245e21f754\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8a671718eeac7e7a27a1d07d50454c998d74d569c4124f8420d70d2cf9c45f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dae05ec584cf62f73c559c6fa11db8094b79f114cb3cfda0476a51e94a2b2c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://17e5ff1188b4d038cf297a11e264b07f37009dda7dbdc9e238b8f2540a13cfd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6427cf9d1469a37e0911c03afa54e7434b985915eeff09153f7208570698af18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a717b3839d629c3c53f54d2805a3fdd9c133ed7183a4169e2e3d40efc712a3b8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-spxq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vbjz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.954174 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"addb758f-1f34-4793-af67-1a54167543b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01a24de3733e43edc33f6dc2fdd238ea7d68abed51646d721614de97674c7576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqx62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-x6v7p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.957793 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.957820 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.957831 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.957861 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.957869 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:37Z","lastTransitionTime":"2025-10-10T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.968273 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88d75259e3e973a1cd2f2a275f262d6547ad03da5e2a9774dda3bd2434688b8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:37 crc kubenswrapper[4669]: I1010 09:12:37.980344 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:37Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.006813 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T09:12:22Z\\\",\\\"message\\\":\\\"ster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561279 6679 services_controller.go:452] Built service openshift-kube-scheduler-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561286 6679 services_controller.go:453] Built service openshift-kube-scheduler-operator/metrics template LB for network=default: []services.LB{}\\\\nI1010 09:12:22.561293 6679 services_controller.go:454] Service openshift-kube-scheduler-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1010 09:12:22.561308 6679 kube.go:317] Updating pod openshift-multus/network-metrics-daemon-bhn82\\\\nI1010 09:12:22.561304 6679 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"1dc899db-4498-4b7a-8437-861940b962e7\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-scheduler-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T09:12:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6xsrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gbsxj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.015953 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-922g5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4486ac07-abf2-46bf-b5d3-354bde8d3b10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0201fd8840235a118f8b48ef56d50699257a08a857cb583fb3fa568d03f575ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b8t7n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:11:17Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-922g5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.027257 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb223aa623615696069a96657cfbdedac3d410b56b2609993e7cc2437f5ba84e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8af685f56fe39646cfb46b9b77b82ae02c80eaa18189bae69b1df8c458d187dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.041228 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d64e629-1220-4f7a-89f3-103f391166bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc204570df4adc06a02fc3b1ab77cd5190f92c18d34352c90237b1a24537aa7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://098612023f457d0f0a9bc6fc7781e82d55279ad99d1ab503fbb2a3942df45d07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac75de1e1009f71e2259adf059abcb4a4113d17d9a42242f39e54f83317b9940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8687746d1a75bde3a0e38f482f89f01385819a02a6806fe3f0bb8d018dc4fb86\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.056324 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cea29cbd-21a1-4963-a294-c755b46dfc7a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T09:10:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bee833a094937674acd79d293f8ced8d36ab7a541fbdabba927bf9224cf773bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a13e715d9e31a46304635f7f7db3a399de785d109205d4c4732994adba30cf0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://42396b4abe4f627f6108dc4cddf11806252bae157a2c9a5d39acbc5a3b301beb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T09:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4745cc69d18e6bd67171b4a07ae76ae30313ae21cb47165bc7a8249b7147652f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T09:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T09:10:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T09:10:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.061837 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.061873 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.061885 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.061903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.061915 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.069246 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.082011 4669 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T09:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T09:12:38Z is after 2025-08-24T17:21:41Z" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.164691 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.164751 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.164774 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.164801 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.164821 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.267283 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.267339 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.267361 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.267388 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.267409 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.369650 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.369709 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.369721 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.369736 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.369746 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.472661 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.472708 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.472724 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.472747 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.472764 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.575520 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.575550 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.575558 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.575570 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.575580 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.678757 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.678794 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.678805 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.678821 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.678832 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.781857 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.781896 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.781906 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.781922 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.781933 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.794701 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:38 crc kubenswrapper[4669]: E1010 09:12:38.794836 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.885472 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.885548 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.885572 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.885633 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.885659 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.989392 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.989455 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.989477 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.989505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:38 crc kubenswrapper[4669]: I1010 09:12:38.989525 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:38Z","lastTransitionTime":"2025-10-10T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.092702 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.092791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.092810 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.092832 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.092848 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.196255 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.196320 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.196337 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.196364 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.196387 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.299101 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.299169 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.299196 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.299224 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.299250 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.401760 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.401833 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.401852 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.401874 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.401889 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.504810 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.504871 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.504889 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.504913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.504930 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.607969 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.608156 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.608179 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.608209 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.608235 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.711421 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.711449 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.711457 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.711470 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.711478 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.794924 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.795111 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:39 crc kubenswrapper[4669]: E1010 09:12:39.795261 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.795449 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:39 crc kubenswrapper[4669]: E1010 09:12:39.795627 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:39 crc kubenswrapper[4669]: E1010 09:12:39.795812 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.813833 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.813879 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.813892 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.813909 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.813920 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.917052 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.917117 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.917138 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.917162 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:39 crc kubenswrapper[4669]: I1010 09:12:39.917179 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:39Z","lastTransitionTime":"2025-10-10T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.020676 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.020737 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.020746 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.020762 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.020775 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.124220 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.124258 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.124268 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.124285 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.124295 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.228124 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.228686 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.229231 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.229513 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.229789 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.332385 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.332772 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.332856 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.332951 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.333030 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.436380 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.436409 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.436419 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.436435 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.436445 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.539167 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.539228 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.539244 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.539269 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.539298 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.642800 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.643254 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.643420 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.643562 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.643758 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.747035 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.747072 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.747084 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.747100 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.747111 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.795425 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:40 crc kubenswrapper[4669]: E1010 09:12:40.795691 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.850505 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.850576 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.850602 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.850621 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.850633 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.953340 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.953418 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.953438 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.953466 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:40 crc kubenswrapper[4669]: I1010 09:12:40.953484 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:40Z","lastTransitionTime":"2025-10-10T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.055506 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.055542 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.055553 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.055569 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.055606 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.157985 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.158027 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.158040 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.158057 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.158067 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.261019 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.261089 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.261106 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.261130 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.261151 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.364348 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.364389 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.364401 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.364419 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.364429 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.466725 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.466791 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.466808 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.466834 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.466853 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.570108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.570429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.570500 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.570573 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.570655 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.672801 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.672841 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.672850 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.672863 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.672872 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.775495 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.775538 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.775634 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.775654 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.775667 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.795093 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:41 crc kubenswrapper[4669]: E1010 09:12:41.795191 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.795211 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:41 crc kubenswrapper[4669]: E1010 09:12:41.795311 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.795490 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:41 crc kubenswrapper[4669]: E1010 09:12:41.795765 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.878256 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.878315 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.878336 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.878365 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.878386 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.981152 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.981194 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.981205 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.981221 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:41 crc kubenswrapper[4669]: I1010 09:12:41.981232 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:41Z","lastTransitionTime":"2025-10-10T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.083986 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.084030 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.084040 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.084054 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.084063 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.187115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.187167 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.187181 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.187203 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.187215 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.289808 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.289848 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.289857 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.289872 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.289883 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.393171 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.393243 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.393265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.393293 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.393316 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.496469 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.496546 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.496563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.496617 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.496640 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.601433 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.601518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.601543 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.601570 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.601620 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.704383 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.704449 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.704485 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.704513 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.704533 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.795341 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:42 crc kubenswrapper[4669]: E1010 09:12:42.795509 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.807500 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.807647 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.807667 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.807687 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.807702 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.910909 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.910966 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.910982 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.911006 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:42 crc kubenswrapper[4669]: I1010 09:12:42.911024 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:42Z","lastTransitionTime":"2025-10-10T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.014775 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.014839 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.014862 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.014890 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.014910 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.118464 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.118523 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.118542 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.118568 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.118637 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.221429 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.221515 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.221539 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.221568 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.221617 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.324268 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.324313 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.324327 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.324347 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.324358 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.426473 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.426509 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.426519 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.426535 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.426547 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.530222 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.530257 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.530265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.530279 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.530288 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.632973 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.633462 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.633490 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.633518 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.633539 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.737034 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.737108 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.737126 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.737148 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.737164 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.795212 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:43 crc kubenswrapper[4669]: E1010 09:12:43.795428 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.795459 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.795462 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:43 crc kubenswrapper[4669]: E1010 09:12:43.795528 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:43 crc kubenswrapper[4669]: E1010 09:12:43.795630 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.840916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.840997 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.841020 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.841044 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.841062 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.943913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.943969 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.943986 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.944007 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:43 crc kubenswrapper[4669]: I1010 09:12:43.944020 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:43Z","lastTransitionTime":"2025-10-10T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.046891 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.046949 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.046962 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.046975 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.046984 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.149037 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.149076 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.149086 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.149101 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.149109 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.253014 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.253075 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.253089 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.253115 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.253131 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.357563 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.357708 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.357735 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.357765 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.357794 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.460892 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.460938 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.460953 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.460970 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.460987 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.563276 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.563309 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.563320 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.563335 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.563344 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.666733 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.666799 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.666823 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.666851 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.666874 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.770063 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.770151 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.770175 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.770202 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.770222 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.795110 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:44 crc kubenswrapper[4669]: E1010 09:12:44.795377 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.872575 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.872717 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.872742 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.872772 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.872794 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.975550 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.975624 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.975641 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.975664 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:44 crc kubenswrapper[4669]: I1010 09:12:44.975694 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:44Z","lastTransitionTime":"2025-10-10T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.078367 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.078425 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.078473 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.078501 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.078518 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.180819 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.180895 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.180913 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.180939 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.180957 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.283820 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.283916 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.283936 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.283961 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.283979 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.387179 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.387265 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.387290 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.387325 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.387351 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.491788 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.491877 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.491903 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.491935 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.492125 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.595079 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.595154 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.595175 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.595211 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.595230 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.697798 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.697834 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.697846 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.697863 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.697875 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.795078 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:45 crc kubenswrapper[4669]: E1010 09:12:45.795286 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.795543 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.795704 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:45 crc kubenswrapper[4669]: E1010 09:12:45.795847 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:45 crc kubenswrapper[4669]: E1010 09:12:45.796008 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.801112 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.801157 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.801173 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.801194 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.801211 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.904208 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.904274 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.904303 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.904338 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.904359 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.982651 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.982972 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.983012 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.983037 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:45 crc kubenswrapper[4669]: I1010 09:12:45.983057 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:45Z","lastTransitionTime":"2025-10-10T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.018311 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.018341 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.018352 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.018368 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.018378 4669 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T09:12:46Z","lastTransitionTime":"2025-10-10T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.054367 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc"] Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.054746 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.058544 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.058790 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.058903 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.059062 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.075677 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-922g5" podStartSLOduration=89.075654717 podStartE2EDuration="1m29.075654717s" podCreationTimestamp="2025-10-10 09:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.074154301 +0000 UTC m=+109.090173083" watchObservedRunningTime="2025-10-10 09:12:46.075654717 +0000 UTC m=+109.091673469" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.121216 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.121199621 podStartE2EDuration="52.121199621s" podCreationTimestamp="2025-10-10 09:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.121097487 +0000 UTC m=+109.137116249" watchObservedRunningTime="2025-10-10 09:12:46.121199621 +0000 UTC m=+109.137218363" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.121336 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=87.121332055 podStartE2EDuration="1m27.121332055s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.10691898 +0000 UTC m=+109.122937732" watchObservedRunningTime="2025-10-10 09:12:46.121332055 +0000 UTC m=+109.137350797" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.172095 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e08739e4-8234-4af2-8bb7-8442cc3473fd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.172164 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e08739e4-8234-4af2-8bb7-8442cc3473fd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.172189 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e08739e4-8234-4af2-8bb7-8442cc3473fd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.172284 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08739e4-8234-4af2-8bb7-8442cc3473fd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.172352 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08739e4-8234-4af2-8bb7-8442cc3473fd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.191807 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.191788917 podStartE2EDuration="1m29.191788917s" podCreationTimestamp="2025-10-10 09:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.162024559 +0000 UTC m=+109.178043311" watchObservedRunningTime="2025-10-10 09:12:46.191788917 +0000 UTC m=+109.207807659" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.192099 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=27.192094366 podStartE2EDuration="27.192094366s" podCreationTimestamp="2025-10-10 09:12:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.191538259 +0000 UTC m=+109.207557001" watchObservedRunningTime="2025-10-10 09:12:46.192094366 +0000 UTC m=+109.208113108" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.206426 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-nq6jn" podStartSLOduration=87.206407307 podStartE2EDuration="1m27.206407307s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.206351916 +0000 UTC m=+109.222370658" watchObservedRunningTime="2025-10-10 09:12:46.206407307 +0000 UTC m=+109.222426059" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.226467 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-pvhp4" podStartSLOduration=87.226448484 podStartE2EDuration="1m27.226448484s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.225824676 +0000 UTC m=+109.241843418" watchObservedRunningTime="2025-10-10 09:12:46.226448484 +0000 UTC m=+109.242467236" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.239197 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2v44p" podStartSLOduration=87.239180087 podStartE2EDuration="1m27.239180087s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.238768165 +0000 UTC m=+109.254786917" watchObservedRunningTime="2025-10-10 09:12:46.239180087 +0000 UTC m=+109.255198839" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.261310 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=39.261293769 podStartE2EDuration="39.261293769s" podCreationTimestamp="2025-10-10 09:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.260214965 +0000 UTC m=+109.276233717" watchObservedRunningTime="2025-10-10 09:12:46.261293769 +0000 UTC m=+109.277312521" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.273765 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08739e4-8234-4af2-8bb7-8442cc3473fd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.273829 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e08739e4-8234-4af2-8bb7-8442cc3473fd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.273885 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e08739e4-8234-4af2-8bb7-8442cc3473fd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.273907 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e08739e4-8234-4af2-8bb7-8442cc3473fd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.273931 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08739e4-8234-4af2-8bb7-8442cc3473fd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.274412 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e08739e4-8234-4af2-8bb7-8442cc3473fd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.274582 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e08739e4-8234-4af2-8bb7-8442cc3473fd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.275028 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e08739e4-8234-4af2-8bb7-8442cc3473fd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.282169 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e08739e4-8234-4af2-8bb7-8442cc3473fd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.292319 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08739e4-8234-4af2-8bb7-8442cc3473fd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xw2zc\" (UID: \"e08739e4-8234-4af2-8bb7-8442cc3473fd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.321340 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vbjz8" podStartSLOduration=87.32132498 podStartE2EDuration="1m27.32132498s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.30901775 +0000 UTC m=+109.325036492" watchObservedRunningTime="2025-10-10 09:12:46.32132498 +0000 UTC m=+109.337343722" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.344438 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podStartSLOduration=87.344421871 podStartE2EDuration="1m27.344421871s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:46.321858685 +0000 UTC m=+109.337877427" watchObservedRunningTime="2025-10-10 09:12:46.344421871 +0000 UTC m=+109.360440613" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.377029 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.422644 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" event={"ID":"e08739e4-8234-4af2-8bb7-8442cc3473fd","Type":"ContainerStarted","Data":"926629464a1ada9e4ebb18983d3faf1920da57acee3c71fcb6f0bc4e151db745"} Oct 10 09:12:46 crc kubenswrapper[4669]: I1010 09:12:46.794399 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:46 crc kubenswrapper[4669]: E1010 09:12:46.794832 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:47 crc kubenswrapper[4669]: I1010 09:12:47.426991 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" event={"ID":"e08739e4-8234-4af2-8bb7-8442cc3473fd","Type":"ContainerStarted","Data":"625b7192b0488f32f7ad208f9e9583e90f95e285108e0fa668992eefd0d47bb5"} Oct 10 09:12:47 crc kubenswrapper[4669]: I1010 09:12:47.794393 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:47 crc kubenswrapper[4669]: I1010 09:12:47.794430 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:47 crc kubenswrapper[4669]: I1010 09:12:47.794392 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:47 crc kubenswrapper[4669]: E1010 09:12:47.795612 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:47 crc kubenswrapper[4669]: E1010 09:12:47.795936 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:47 crc kubenswrapper[4669]: E1010 09:12:47.795858 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:48 crc kubenswrapper[4669]: I1010 09:12:48.794738 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:48 crc kubenswrapper[4669]: E1010 09:12:48.794876 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:49 crc kubenswrapper[4669]: I1010 09:12:49.794765 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:49 crc kubenswrapper[4669]: I1010 09:12:49.794916 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:49 crc kubenswrapper[4669]: E1010 09:12:49.795254 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:49 crc kubenswrapper[4669]: I1010 09:12:49.795300 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:49 crc kubenswrapper[4669]: E1010 09:12:49.795502 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:49 crc kubenswrapper[4669]: E1010 09:12:49.796342 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:49 crc kubenswrapper[4669]: I1010 09:12:49.796846 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:12:49 crc kubenswrapper[4669]: E1010 09:12:49.797146 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gbsxj_openshift-ovn-kubernetes(e1b02e9a-7e2e-473d-a810-d4ece0d3a18e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" Oct 10 09:12:50 crc kubenswrapper[4669]: I1010 09:12:50.794359 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:50 crc kubenswrapper[4669]: E1010 09:12:50.794558 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:51 crc kubenswrapper[4669]: I1010 09:12:51.795200 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:51 crc kubenswrapper[4669]: I1010 09:12:51.795232 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:51 crc kubenswrapper[4669]: E1010 09:12:51.795386 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:51 crc kubenswrapper[4669]: I1010 09:12:51.795424 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:51 crc kubenswrapper[4669]: E1010 09:12:51.795566 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:51 crc kubenswrapper[4669]: E1010 09:12:51.795710 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:52 crc kubenswrapper[4669]: I1010 09:12:52.795462 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:52 crc kubenswrapper[4669]: E1010 09:12:52.795706 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.458945 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/1.log" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.459901 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/0.log" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.459951 4669 generic.go:334] "Generic (PLEG): container finished" podID="9502d677-e546-4df7-96c2-bce8274c0f57" containerID="7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7" exitCode=1 Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.459986 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerDied","Data":"7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7"} Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.460024 4669 scope.go:117] "RemoveContainer" containerID="79368b05500bee5b90f51c6eb91692274ee0e206b2d9147cc5108d3e97bda620" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.460472 4669 scope.go:117] "RemoveContainer" containerID="7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7" Oct 10 09:12:53 crc kubenswrapper[4669]: E1010 09:12:53.460659 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-nq6jn_openshift-multus(9502d677-e546-4df7-96c2-bce8274c0f57)\"" pod="openshift-multus/multus-nq6jn" podUID="9502d677-e546-4df7-96c2-bce8274c0f57" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.490219 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xw2zc" podStartSLOduration=94.490194229 podStartE2EDuration="1m34.490194229s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:12:47.444131359 +0000 UTC m=+110.460150111" watchObservedRunningTime="2025-10-10 09:12:53.490194229 +0000 UTC m=+116.506213001" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.794942 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.794986 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:53 crc kubenswrapper[4669]: E1010 09:12:53.795096 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:53 crc kubenswrapper[4669]: E1010 09:12:53.795243 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:53 crc kubenswrapper[4669]: I1010 09:12:53.794942 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:53 crc kubenswrapper[4669]: E1010 09:12:53.795352 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:54 crc kubenswrapper[4669]: I1010 09:12:54.465683 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/1.log" Oct 10 09:12:54 crc kubenswrapper[4669]: I1010 09:12:54.795328 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:54 crc kubenswrapper[4669]: E1010 09:12:54.795513 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:55 crc kubenswrapper[4669]: I1010 09:12:55.795079 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:55 crc kubenswrapper[4669]: I1010 09:12:55.795171 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:55 crc kubenswrapper[4669]: I1010 09:12:55.795167 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:55 crc kubenswrapper[4669]: E1010 09:12:55.795327 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:55 crc kubenswrapper[4669]: E1010 09:12:55.795493 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:55 crc kubenswrapper[4669]: E1010 09:12:55.795718 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:56 crc kubenswrapper[4669]: I1010 09:12:56.794921 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:56 crc kubenswrapper[4669]: E1010 09:12:56.795280 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:57 crc kubenswrapper[4669]: E1010 09:12:57.756652 4669 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 10 09:12:57 crc kubenswrapper[4669]: I1010 09:12:57.795051 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:57 crc kubenswrapper[4669]: I1010 09:12:57.795104 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:57 crc kubenswrapper[4669]: I1010 09:12:57.795165 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:57 crc kubenswrapper[4669]: E1010 09:12:57.797013 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:57 crc kubenswrapper[4669]: E1010 09:12:57.797285 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:12:57 crc kubenswrapper[4669]: E1010 09:12:57.797406 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:57 crc kubenswrapper[4669]: E1010 09:12:57.884649 4669 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 10 09:12:58 crc kubenswrapper[4669]: I1010 09:12:58.795217 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:12:58 crc kubenswrapper[4669]: E1010 09:12:58.796197 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:12:59 crc kubenswrapper[4669]: I1010 09:12:59.794463 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:12:59 crc kubenswrapper[4669]: E1010 09:12:59.794750 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:12:59 crc kubenswrapper[4669]: I1010 09:12:59.794857 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:12:59 crc kubenswrapper[4669]: E1010 09:12:59.794977 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:12:59 crc kubenswrapper[4669]: I1010 09:12:59.794851 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:12:59 crc kubenswrapper[4669]: E1010 09:12:59.795145 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:13:00 crc kubenswrapper[4669]: I1010 09:13:00.795495 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:00 crc kubenswrapper[4669]: E1010 09:13:00.796863 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:13:01 crc kubenswrapper[4669]: I1010 09:13:01.794791 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:01 crc kubenswrapper[4669]: I1010 09:13:01.794822 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:01 crc kubenswrapper[4669]: E1010 09:13:01.795100 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:13:01 crc kubenswrapper[4669]: I1010 09:13:01.795190 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:01 crc kubenswrapper[4669]: E1010 09:13:01.795363 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:13:01 crc kubenswrapper[4669]: E1010 09:13:01.795520 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:13:02 crc kubenswrapper[4669]: I1010 09:13:02.795017 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:02 crc kubenswrapper[4669]: E1010 09:13:02.795210 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:13:02 crc kubenswrapper[4669]: E1010 09:13:02.886010 4669 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 10 09:13:03 crc kubenswrapper[4669]: I1010 09:13:03.795518 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:03 crc kubenswrapper[4669]: I1010 09:13:03.795571 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:03 crc kubenswrapper[4669]: I1010 09:13:03.795753 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:03 crc kubenswrapper[4669]: E1010 09:13:03.795769 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:13:03 crc kubenswrapper[4669]: E1010 09:13:03.795859 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:13:03 crc kubenswrapper[4669]: I1010 09:13:03.796462 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:13:03 crc kubenswrapper[4669]: E1010 09:13:03.796851 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:13:04 crc kubenswrapper[4669]: I1010 09:13:04.504859 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/3.log" Oct 10 09:13:04 crc kubenswrapper[4669]: I1010 09:13:04.507423 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerStarted","Data":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} Oct 10 09:13:04 crc kubenswrapper[4669]: I1010 09:13:04.507823 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:13:04 crc kubenswrapper[4669]: I1010 09:13:04.534043 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podStartSLOduration=105.53402425 podStartE2EDuration="1m45.53402425s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:04.533138442 +0000 UTC m=+127.549157204" watchObservedRunningTime="2025-10-10 09:13:04.53402425 +0000 UTC m=+127.550042992" Oct 10 09:13:04 crc kubenswrapper[4669]: I1010 09:13:04.730032 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-bhn82"] Oct 10 09:13:04 crc kubenswrapper[4669]: I1010 09:13:04.730237 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:04 crc kubenswrapper[4669]: E1010 09:13:04.730453 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:13:04 crc kubenswrapper[4669]: I1010 09:13:04.794908 4669 scope.go:117] "RemoveContainer" containerID="7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7" Oct 10 09:13:05 crc kubenswrapper[4669]: I1010 09:13:05.512866 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/1.log" Oct 10 09:13:05 crc kubenswrapper[4669]: I1010 09:13:05.513271 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerStarted","Data":"7b2f37ba6882cf12f83ece0045d1343775e215b0d1ebf073935817d0e74dc426"} Oct 10 09:13:05 crc kubenswrapper[4669]: I1010 09:13:05.795164 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:05 crc kubenswrapper[4669]: I1010 09:13:05.795327 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:05 crc kubenswrapper[4669]: E1010 09:13:05.795972 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:13:05 crc kubenswrapper[4669]: E1010 09:13:05.796159 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:13:05 crc kubenswrapper[4669]: I1010 09:13:05.796339 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:05 crc kubenswrapper[4669]: E1010 09:13:05.796628 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:13:06 crc kubenswrapper[4669]: I1010 09:13:06.795444 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:06 crc kubenswrapper[4669]: E1010 09:13:06.795750 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bhn82" podUID="20d5735d-7ca2-4824-9b5d-4bb39502a3dc" Oct 10 09:13:07 crc kubenswrapper[4669]: I1010 09:13:07.795048 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:07 crc kubenswrapper[4669]: I1010 09:13:07.795084 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:07 crc kubenswrapper[4669]: E1010 09:13:07.797307 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 09:13:07 crc kubenswrapper[4669]: I1010 09:13:07.797438 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:07 crc kubenswrapper[4669]: E1010 09:13:07.797558 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 09:13:07 crc kubenswrapper[4669]: E1010 09:13:07.797670 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 09:13:08 crc kubenswrapper[4669]: I1010 09:13:08.794843 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:08 crc kubenswrapper[4669]: I1010 09:13:08.798909 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 10 09:13:08 crc kubenswrapper[4669]: I1010 09:13:08.799016 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 10 09:13:09 crc kubenswrapper[4669]: I1010 09:13:09.795345 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:09 crc kubenswrapper[4669]: I1010 09:13:09.795382 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:09 crc kubenswrapper[4669]: I1010 09:13:09.795622 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:09 crc kubenswrapper[4669]: I1010 09:13:09.798393 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 10 09:13:09 crc kubenswrapper[4669]: I1010 09:13:09.799002 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 10 09:13:09 crc kubenswrapper[4669]: I1010 09:13:09.799179 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 10 09:13:09 crc kubenswrapper[4669]: I1010 09:13:09.800106 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.635654 4669 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.686559 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-x6svn"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.687633 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wqc7t"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.687680 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.688174 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.691953 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.692747 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.699060 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 10 09:13:16 crc kubenswrapper[4669]: W1010 09:13:16.699188 4669 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 10 09:13:16 crc kubenswrapper[4669]: E1010 09:13:16.699257 4669 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:13:16 crc kubenswrapper[4669]: W1010 09:13:16.699370 4669 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 10 09:13:16 crc kubenswrapper[4669]: E1010 09:13:16.699401 4669 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:13:16 crc kubenswrapper[4669]: W1010 09:13:16.700762 4669 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.700821 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 10 09:13:16 crc kubenswrapper[4669]: E1010 09:13:16.700829 4669 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:13:16 crc kubenswrapper[4669]: W1010 09:13:16.700769 4669 reflector.go:561] object-"openshift-route-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 10 09:13:16 crc kubenswrapper[4669]: E1010 09:13:16.700892 4669 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.707603 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.707971 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.708111 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 10 09:13:16 crc kubenswrapper[4669]: W1010 09:13:16.708241 4669 reflector.go:561] object-"openshift-route-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 10 09:13:16 crc kubenswrapper[4669]: E1010 09:13:16.708278 4669 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.708455 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.708638 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.709089 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.709513 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.710007 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-oauth-config\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728408 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-console-config\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728558 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff6pg\" (UniqueName: \"kubernetes.io/projected/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-kube-api-access-ff6pg\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728647 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shld6\" (UniqueName: \"kubernetes.io/projected/69730259-7614-4869-bc84-10e17d8d13df-kube-api-access-shld6\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728701 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-etcd-client\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728745 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-audit-dir\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728821 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728925 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-oauth-serving-cert\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.728968 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-encryption-config\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729012 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-config\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729054 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-node-pullsecrets\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729107 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-audit\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729142 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-serving-cert\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729183 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-service-ca\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729261 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-image-import-ca\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729293 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-trusted-ca-bundle\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729364 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-etcd-serving-ca\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.729413 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-serving-cert\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.710357 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.710470 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.710554 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.710685 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 10 09:13:16 crc kubenswrapper[4669]: W1010 09:13:16.710751 4669 reflector.go:561] object-"openshift-route-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 10 09:13:16 crc kubenswrapper[4669]: E1010 09:13:16.730208 4669 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.711076 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.711890 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.714728 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.711974 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.712163 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.715766 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.732347 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9lpfb"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.732984 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.733998 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.735227 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.736051 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.750621 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ngvr9"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.770568 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.771044 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.771463 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772287 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nbx9n"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772357 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ngvr9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772405 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772645 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772766 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772915 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772947 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772961 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.773059 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.773174 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.773415 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.773643 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.772925 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.774536 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.775321 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.775441 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.775448 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.775465 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.775544 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.775704 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.777877 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.778350 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4mpz6"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.778785 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.779691 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.780787 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.781153 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.783607 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb9vp"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.784154 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rcqcc"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.784743 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.785190 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.785237 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.785400 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.786939 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-cr5vp"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.787678 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.788126 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8q8w9"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.789990 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.790559 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.795141 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.796112 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.796887 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.797253 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.797517 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.797759 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798052 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798249 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798370 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798465 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798601 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798680 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798785 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798842 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.798857 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799051 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799125 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799168 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799233 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799292 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799331 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799052 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799483 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799557 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799680 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799765 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799850 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.799932 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.800055 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.800159 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.800058 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nvfcj"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.800424 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.800798 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.801940 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.802335 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.802549 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.802905 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.804719 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.805263 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.805610 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.805681 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.805900 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.806190 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.806676 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.810439 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.811675 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.811844 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838045 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838324 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmcpz\" (UniqueName: \"kubernetes.io/projected/52821f37-1b46-4913-b0ac-c00e0a899193-kube-api-access-cmcpz\") pod \"cluster-samples-operator-665b6dd947-j89vb\" (UID: \"52821f37-1b46-4913-b0ac-c00e0a899193\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838358 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e57d53b2-5487-4820-aa61-9070eb866eae-images\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838399 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-serving-cert\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838420 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnk4t\" (UniqueName: \"kubernetes.io/projected/182c02d0-fdb0-4a01-bb6f-efdc3bbcb550-kube-api-access-qnk4t\") pod \"multus-admission-controller-857f4d67dd-rcqcc\" (UID: \"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838441 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxz7t\" (UniqueName: \"kubernetes.io/projected/55a8f8c1-7a6c-474f-b422-60d886882a59-kube-api-access-xxz7t\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838475 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-encryption-config\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838496 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-dir\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838516 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf8wq\" (UniqueName: \"kubernetes.io/projected/eb510d78-5e60-48d3-8deb-3e7386845e0a-kube-api-access-gf8wq\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838551 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-service-ca\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838568 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838608 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-auth-proxy-config\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.838628 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-signing-key\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839153 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjcth\" (UniqueName: \"kubernetes.io/projected/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-kube-api-access-vjcth\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839252 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-etcd-client\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839345 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839369 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55a8f8c1-7a6c-474f-b422-60d886882a59-serving-cert\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839459 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-serving-cert\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839594 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-image-import-ca\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839633 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-trusted-ca-bundle\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839668 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3df4b15-909b-4a5d-af3a-af282d2ee197-images\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839686 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24dc9ac3-3322-40f6-a99b-433280694265-serving-cert\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839705 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk2dk\" (UniqueName: \"kubernetes.io/projected/24dc9ac3-3322-40f6-a99b-433280694265-kube-api-access-zk2dk\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839736 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls7nt\" (UniqueName: \"kubernetes.io/projected/3f266776-152e-4b8c-b4e4-98a578b5f238-kube-api-access-ls7nt\") pod \"downloads-7954f5f757-ngvr9\" (UID: \"3f266776-152e-4b8c-b4e4-98a578b5f238\") " pod="openshift-console/downloads-7954f5f757-ngvr9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839753 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3df4b15-909b-4a5d-af3a-af282d2ee197-proxy-tls\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839786 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-etcd-serving-ca\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.839857 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-serving-cert\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.844779 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fcf6de-fdb5-4759-9b09-01e59eca2655-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.844826 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.844861 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-serving-cert\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.844891 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-signing-cabundle\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.844918 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fcf6de-fdb5-4759-9b09-01e59eca2655-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.844948 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-config\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.844984 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-oauth-config\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845009 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-console-config\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845035 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e57d53b2-5487-4820-aa61-9070eb866eae-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845061 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpxfm\" (UniqueName: \"kubernetes.io/projected/e57d53b2-5487-4820-aa61-9070eb866eae-kube-api-access-xpxfm\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845081 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq9qk\" (UniqueName: \"kubernetes.io/projected/0342148d-ffe9-4d1d-a629-03108f4c2cda-kube-api-access-kq9qk\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845105 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/55a8f8c1-7a6c-474f-b422-60d886882a59-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845129 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm76k\" (UniqueName: \"kubernetes.io/projected/d3df4b15-909b-4a5d-af3a-af282d2ee197-kube-api-access-zm76k\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845151 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845170 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggp22\" (UniqueName: \"kubernetes.io/projected/2a92e481-9a5c-4238-988f-f2d0b852ebca-kube-api-access-ggp22\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845192 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e57d53b2-5487-4820-aa61-9070eb866eae-config\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845231 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-584rk\" (UniqueName: \"kubernetes.io/projected/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-kube-api-access-584rk\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845252 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845276 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff6pg\" (UniqueName: \"kubernetes.io/projected/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-kube-api-access-ff6pg\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845296 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845318 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845339 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845360 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shld6\" (UniqueName: \"kubernetes.io/projected/69730259-7614-4869-bc84-10e17d8d13df-kube-api-access-shld6\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845383 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb510d78-5e60-48d3-8deb-3e7386845e0a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845405 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845433 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-etcd-client\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845452 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-audit-dir\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845471 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5962f59-9740-46e8-9866-a57806721845-audit-dir\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845497 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845520 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845538 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-config\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845559 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-audit-policies\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845617 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/182c02d0-fdb0-4a01-bb6f-efdc3bbcb550-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rcqcc\" (UID: \"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845637 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb510d78-5e60-48d3-8deb-3e7386845e0a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845661 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-machine-approver-tls\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845678 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845703 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845723 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845759 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-client-ca\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845800 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845819 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845854 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-config\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845874 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-trusted-ca\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845894 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3df4b15-909b-4a5d-af3a-af282d2ee197-auth-proxy-config\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845912 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-oauth-serving-cert\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845931 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845951 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/52821f37-1b46-4913-b0ac-c00e0a899193-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j89vb\" (UID: \"52821f37-1b46-4913-b0ac-c00e0a899193\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845971 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-service-ca-bundle\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.845992 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-encryption-config\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846013 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjds2\" (UniqueName: \"kubernetes.io/projected/521f5278-6996-4b67-a6ab-d46bd1a35acb-kube-api-access-pjds2\") pod \"migrator-59844c95c7-lvq4s\" (UID: \"521f5278-6996-4b67-a6ab-d46bd1a35acb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846034 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4j5h\" (UniqueName: \"kubernetes.io/projected/99fcf6de-fdb5-4759-9b09-01e59eca2655-kube-api-access-k4j5h\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846052 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846073 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-config\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846090 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9mv2\" (UniqueName: \"kubernetes.io/projected/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-kube-api-access-d9mv2\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846111 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-node-pullsecrets\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846132 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846152 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0342148d-ffe9-4d1d-a629-03108f4c2cda-serving-cert\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846171 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846190 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-policies\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846208 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846234 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6clmf\" (UniqueName: \"kubernetes.io/projected/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-kube-api-access-6clmf\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846255 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsc44\" (UniqueName: \"kubernetes.io/projected/e5962f59-9740-46e8-9866-a57806721845-kube-api-access-wsc44\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846271 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-config\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846266 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-image-import-ca\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.847078 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-etcd-serving-ca\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.847756 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-trusted-ca-bundle\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.846292 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq79w\" (UniqueName: \"kubernetes.io/projected/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-kube-api-access-nq79w\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.847850 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-audit\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.847897 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.841960 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-service-ca\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.853136 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.854204 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-audit\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.854406 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.854977 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.855436 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.856431 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-oauth-serving-cert\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.856618 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-node-pullsecrets\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.864191 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-config\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.864484 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-audit-dir\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.866668 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-serving-cert\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.867241 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.867899 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-oauth-config\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.867932 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.868164 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.868493 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.869707 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-encryption-config\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.870016 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.870477 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.870649 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.870832 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.871133 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.871221 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.871373 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.871436 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.873033 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.873461 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.876915 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-console-config\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.886527 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-etcd-client\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.905393 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-serving-cert\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.906193 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.906342 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.906502 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.906559 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.906831 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.906919 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.906961 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.907506 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.907665 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.907869 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.907954 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.908164 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-d7jk5"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.910927 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.912050 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.914930 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.920535 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.921859 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.922644 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.923045 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.923140 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-klgpb"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.923666 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.924022 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.924246 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.927033 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.929707 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.931908 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.937343 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.937502 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-zqhsj"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.938003 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.938383 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.938567 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.938648 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w8ngm"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.939547 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.941678 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ngvr9"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.941791 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.942851 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.943169 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.944598 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.944849 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.953094 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960739 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf8wq\" (UniqueName: \"kubernetes.io/projected/eb510d78-5e60-48d3-8deb-3e7386845e0a-kube-api-access-gf8wq\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960793 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960839 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-auth-proxy-config\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960868 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-signing-key\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960893 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjcth\" (UniqueName: \"kubernetes.io/projected/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-kube-api-access-vjcth\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960931 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c2151912-a994-4992-915d-d737497335d7-webhook-cert\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960964 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-etcd-client\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.960996 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961026 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55a8f8c1-7a6c-474f-b422-60d886882a59-serving-cert\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961072 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-serving-cert\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961106 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ct8g\" (UniqueName: \"kubernetes.io/projected/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-kube-api-access-7ct8g\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961136 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3df4b15-909b-4a5d-af3a-af282d2ee197-images\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961163 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961196 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24dc9ac3-3322-40f6-a99b-433280694265-serving-cert\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961225 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk2dk\" (UniqueName: \"kubernetes.io/projected/24dc9ac3-3322-40f6-a99b-433280694265-kube-api-access-zk2dk\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961252 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3df4b15-909b-4a5d-af3a-af282d2ee197-proxy-tls\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961297 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls7nt\" (UniqueName: \"kubernetes.io/projected/3f266776-152e-4b8c-b4e4-98a578b5f238-kube-api-access-ls7nt\") pod \"downloads-7954f5f757-ngvr9\" (UID: \"3f266776-152e-4b8c-b4e4-98a578b5f238\") " pod="openshift-console/downloads-7954f5f757-ngvr9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961323 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fcf6de-fdb5-4759-9b09-01e59eca2655-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961353 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-serving-cert\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961383 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-signing-cabundle\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961407 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fcf6de-fdb5-4759-9b09-01e59eca2655-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961434 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m2ql\" (UniqueName: \"kubernetes.io/projected/89852882-ff51-4b3c-9e60-8a862b7d18cc-kube-api-access-7m2ql\") pod \"package-server-manager-789f6589d5-bngrg\" (UID: \"89852882-ff51-4b3c-9e60-8a862b7d18cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961459 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961488 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-config\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961513 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e57d53b2-5487-4820-aa61-9070eb866eae-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961539 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efe68d-5663-488a-92ca-c02a288b524f-config-volume\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961575 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c2151912-a994-4992-915d-d737497335d7-apiservice-cert\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961620 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq9qk\" (UniqueName: \"kubernetes.io/projected/0342148d-ffe9-4d1d-a629-03108f4c2cda-kube-api-access-kq9qk\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961646 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/55a8f8c1-7a6c-474f-b422-60d886882a59-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961671 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm76k\" (UniqueName: \"kubernetes.io/projected/d3df4b15-909b-4a5d-af3a-af282d2ee197-kube-api-access-zm76k\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961696 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpxfm\" (UniqueName: \"kubernetes.io/projected/e57d53b2-5487-4820-aa61-9070eb866eae-kube-api-access-xpxfm\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961724 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961736 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-auth-proxy-config\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961753 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggp22\" (UniqueName: \"kubernetes.io/projected/2a92e481-9a5c-4238-988f-f2d0b852ebca-kube-api-access-ggp22\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961782 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e57d53b2-5487-4820-aa61-9070eb866eae-config\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961811 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efe68d-5663-488a-92ca-c02a288b524f-secret-volume\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961851 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q62c6\" (UniqueName: \"kubernetes.io/projected/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-kube-api-access-q62c6\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961878 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961908 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961938 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-584rk\" (UniqueName: \"kubernetes.io/projected/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-kube-api-access-584rk\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961973 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.961997 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962020 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962053 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb510d78-5e60-48d3-8deb-3e7386845e0a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962080 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962106 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962139 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5962f59-9740-46e8-9866-a57806721845-audit-dir\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962171 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962198 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962220 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-config\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962246 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f273bc2e-26d7-40fe-9150-fb582fb14618-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-f6dtf\" (UID: \"f273bc2e-26d7-40fe-9150-fb582fb14618\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962274 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c2151912-a994-4992-915d-d737497335d7-tmpfs\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962302 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/acbd17d0-d2e2-4a68-991c-3553eb69f8ea-metrics-tls\") pod \"dns-operator-744455d44c-d7jk5\" (UID: \"acbd17d0-d2e2-4a68-991c-3553eb69f8ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962324 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-audit-policies\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962347 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962378 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/182c02d0-fdb0-4a01-bb6f-efdc3bbcb550-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rcqcc\" (UID: \"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962403 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb510d78-5e60-48d3-8deb-3e7386845e0a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962430 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-machine-approver-tls\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962452 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962484 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/89852882-ff51-4b3c-9e60-8a862b7d18cc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bngrg\" (UID: \"89852882-ff51-4b3c-9e60-8a862b7d18cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962511 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962537 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-client-ca\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962597 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962625 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962659 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-config\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962682 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jffg9\" (UniqueName: \"kubernetes.io/projected/98efe68d-5663-488a-92ca-c02a288b524f-kube-api-access-jffg9\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962704 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3df4b15-909b-4a5d-af3a-af282d2ee197-auth-proxy-config\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962729 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8v7p\" (UniqueName: \"kubernetes.io/projected/c2151912-a994-4992-915d-d737497335d7-kube-api-access-f8v7p\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962752 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962779 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-trusted-ca\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962810 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-service-ca-bundle\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962835 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/52821f37-1b46-4913-b0ac-c00e0a899193-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j89vb\" (UID: \"52821f37-1b46-4913-b0ac-c00e0a899193\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962860 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjds2\" (UniqueName: \"kubernetes.io/projected/521f5278-6996-4b67-a6ab-d46bd1a35acb-kube-api-access-pjds2\") pod \"migrator-59844c95c7-lvq4s\" (UID: \"521f5278-6996-4b67-a6ab-d46bd1a35acb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962888 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4j5h\" (UniqueName: \"kubernetes.io/projected/99fcf6de-fdb5-4759-9b09-01e59eca2655-kube-api-access-k4j5h\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962912 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962942 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrkw9\" (UniqueName: \"kubernetes.io/projected/acbd17d0-d2e2-4a68-991c-3553eb69f8ea-kube-api-access-wrkw9\") pod \"dns-operator-744455d44c-d7jk5\" (UID: \"acbd17d0-d2e2-4a68-991c-3553eb69f8ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962964 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.962991 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9mv2\" (UniqueName: \"kubernetes.io/projected/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-kube-api-access-d9mv2\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963018 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqdcx\" (UniqueName: \"kubernetes.io/projected/15cafda9-1dc9-40b0-af56-024dfaf32754-kube-api-access-wqdcx\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963045 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963068 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0342148d-ffe9-4d1d-a629-03108f4c2cda-serving-cert\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963093 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-policies\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963116 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963142 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/15cafda9-1dc9-40b0-af56-024dfaf32754-srv-cert\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963164 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/15cafda9-1dc9-40b0-af56-024dfaf32754-profile-collector-cert\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963185 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsc44\" (UniqueName: \"kubernetes.io/projected/e5962f59-9740-46e8-9866-a57806721845-kube-api-access-wsc44\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963207 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-config\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963230 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6clmf\" (UniqueName: \"kubernetes.io/projected/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-kube-api-access-6clmf\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963254 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963274 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq79w\" (UniqueName: \"kubernetes.io/projected/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-kube-api-access-nq79w\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963300 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e57d53b2-5487-4820-aa61-9070eb866eae-images\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963331 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnk4t\" (UniqueName: \"kubernetes.io/projected/182c02d0-fdb0-4a01-bb6f-efdc3bbcb550-kube-api-access-qnk4t\") pod \"multus-admission-controller-857f4d67dd-rcqcc\" (UID: \"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963358 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmcpz\" (UniqueName: \"kubernetes.io/projected/52821f37-1b46-4913-b0ac-c00e0a899193-kube-api-access-cmcpz\") pod \"cluster-samples-operator-665b6dd947-j89vb\" (UID: \"52821f37-1b46-4913-b0ac-c00e0a899193\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963386 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkn84\" (UniqueName: \"kubernetes.io/projected/f273bc2e-26d7-40fe-9150-fb582fb14618-kube-api-access-jkn84\") pod \"control-plane-machine-set-operator-78cbb6b69f-f6dtf\" (UID: \"f273bc2e-26d7-40fe-9150-fb582fb14618\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963391 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963407 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-encryption-config\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963436 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963488 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-dir\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.964260 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.964927 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.963435 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-dir\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.966850 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxz7t\" (UniqueName: \"kubernetes.io/projected/55a8f8c1-7a6c-474f-b422-60d886882a59-kube-api-access-xxz7t\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.967408 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.972544 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.973557 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/55a8f8c1-7a6c-474f-b422-60d886882a59-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.977321 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.978559 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.978575 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mfjcd"] Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.980420 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.981607 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e57d53b2-5487-4820-aa61-9070eb866eae-config\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.988859 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.991008 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.991810 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.994222 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.994870 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-policies\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.996087 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.996355 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-config\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.996567 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-config\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.996986 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d3df4b15-909b-4a5d-af3a-af282d2ee197-auth-proxy-config\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.997788 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e57d53b2-5487-4820-aa61-9070eb866eae-images\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.997792 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:16 crc kubenswrapper[4669]: I1010 09:13:16.999809 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-trusted-ca\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.000533 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-service-ca-bundle\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.002385 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-x6svn"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.004021 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.007539 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-etcd-client\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.007857 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/52821f37-1b46-4913-b0ac-c00e0a899193-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j89vb\" (UID: \"52821f37-1b46-4913-b0ac-c00e0a899193\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.007949 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.008442 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb510d78-5e60-48d3-8deb-3e7386845e0a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.009836 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.010080 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.010566 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.011404 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e57d53b2-5487-4820-aa61-9070eb866eae-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.011687 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9lpfb"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.011837 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.012067 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-signing-key\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.012226 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e5962f59-9740-46e8-9866-a57806721845-audit-dir\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.013741 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24dc9ac3-3322-40f6-a99b-433280694265-config\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.017550 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.017620 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.018514 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d3df4b15-909b-4a5d-af3a-af282d2ee197-images\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.020103 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24dc9ac3-3322-40f6-a99b-433280694265-serving-cert\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.036370 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb510d78-5e60-48d3-8deb-3e7386845e0a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.037978 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55a8f8c1-7a6c-474f-b422-60d886882a59-serving-cert\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.038517 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.039186 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.039728 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.039814 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rcqcc"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.040197 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.040388 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-encryption-config\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.040611 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.040668 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.040965 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-serving-cert\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.041442 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8q8w9"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.042139 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.042502 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-signing-cabundle\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.044307 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.045911 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5962f59-9740-46e8-9866-a57806721845-serving-cert\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.046839 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-machine-approver-tls\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.047089 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.047380 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.047544 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.051132 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d3df4b15-909b-4a5d-af3a-af282d2ee197-proxy-tls\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.051198 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.051248 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e5962f59-9740-46e8-9866-a57806721845-audit-policies\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.051808 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4mpz6"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.054323 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/182c02d0-fdb0-4a01-bb6f-efdc3bbcb550-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rcqcc\" (UID: \"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.054603 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.056446 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.059043 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.059102 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fcf6de-fdb5-4759-9b09-01e59eca2655-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.059272 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.060633 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.060699 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb9vp"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.061388 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nbx9n"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.062528 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wqc7t"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.063460 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99fcf6de-fdb5-4759-9b09-01e59eca2655-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.063484 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-config\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.066200 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0342148d-ffe9-4d1d-a629-03108f4c2cda-serving-cert\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.067906 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efe68d-5663-488a-92ca-c02a288b524f-config-volume\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.067928 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c2151912-a994-4992-915d-d737497335d7-apiservice-cert\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.067968 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efe68d-5663-488a-92ca-c02a288b524f-secret-volume\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068007 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068046 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jffg9\" (UniqueName: \"kubernetes.io/projected/98efe68d-5663-488a-92ca-c02a288b524f-kube-api-access-jffg9\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068076 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrkw9\" (UniqueName: \"kubernetes.io/projected/acbd17d0-d2e2-4a68-991c-3553eb69f8ea-kube-api-access-wrkw9\") pod \"dns-operator-744455d44c-d7jk5\" (UID: \"acbd17d0-d2e2-4a68-991c-3553eb69f8ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068099 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/15cafda9-1dc9-40b0-af56-024dfaf32754-profile-collector-cert\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068119 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkn84\" (UniqueName: \"kubernetes.io/projected/f273bc2e-26d7-40fe-9150-fb582fb14618-kube-api-access-jkn84\") pod \"control-plane-machine-set-operator-78cbb6b69f-f6dtf\" (UID: \"f273bc2e-26d7-40fe-9150-fb582fb14618\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068139 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fb25\" (UniqueName: \"kubernetes.io/projected/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-kube-api-access-9fb25\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068164 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ct8g\" (UniqueName: \"kubernetes.io/projected/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-kube-api-access-7ct8g\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068192 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q62c6\" (UniqueName: \"kubernetes.io/projected/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-kube-api-access-q62c6\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068208 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068235 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068256 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f273bc2e-26d7-40fe-9150-fb582fb14618-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-f6dtf\" (UID: \"f273bc2e-26d7-40fe-9150-fb582fb14618\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068272 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c2151912-a994-4992-915d-d737497335d7-tmpfs\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068287 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/acbd17d0-d2e2-4a68-991c-3553eb69f8ea-metrics-tls\") pod \"dns-operator-744455d44c-d7jk5\" (UID: \"acbd17d0-d2e2-4a68-991c-3553eb69f8ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068304 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjprf\" (UniqueName: \"kubernetes.io/projected/c62d6a28-def0-41d5-a931-3bb27183deef-kube-api-access-wjprf\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068322 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c62d6a28-def0-41d5-a931-3bb27183deef-srv-cert\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068340 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/89852882-ff51-4b3c-9e60-8a862b7d18cc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bngrg\" (UID: \"89852882-ff51-4b3c-9e60-8a862b7d18cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068362 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8v7p\" (UniqueName: \"kubernetes.io/projected/c2151912-a994-4992-915d-d737497335d7-kube-api-access-f8v7p\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068397 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c62d6a28-def0-41d5-a931-3bb27183deef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068416 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqdcx\" (UniqueName: \"kubernetes.io/projected/15cafda9-1dc9-40b0-af56-024dfaf32754-kube-api-access-wqdcx\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068446 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/15cafda9-1dc9-40b0-af56-024dfaf32754-srv-cert\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068463 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068512 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c2151912-a994-4992-915d-d737497335d7-webhook-cert\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068530 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-proxy-tls\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068547 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068578 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m2ql\" (UniqueName: \"kubernetes.io/projected/89852882-ff51-4b3c-9e60-8a862b7d18cc-kube-api-access-7m2ql\") pod \"package-server-manager-789f6589d5-bngrg\" (UID: \"89852882-ff51-4b3c-9e60-8a862b7d18cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.068667 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-llrlv"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.069317 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rnpbk"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.069559 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c2151912-a994-4992-915d-d737497335d7-tmpfs\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.069892 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.070142 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.071538 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.074760 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.075154 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mfjcd"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.076819 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-client-ca\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.085342 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-llrlv"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.087944 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-d7jk5"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.089910 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-klgpb"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.091430 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.092754 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-cr5vp"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.098006 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.100417 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.101484 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.102643 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.104695 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nvfcj"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.105863 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.106455 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.107331 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w8ngm"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.108474 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.109795 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.109931 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.111150 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hsgp7"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.111849 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.112098 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.113254 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.114223 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rnpbk"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.123676 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/89852882-ff51-4b3c-9e60-8a862b7d18cc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bngrg\" (UID: \"89852882-ff51-4b3c-9e60-8a862b7d18cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.129474 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.149834 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.162642 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efe68d-5663-488a-92ca-c02a288b524f-secret-volume\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.163269 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/15cafda9-1dc9-40b0-af56-024dfaf32754-profile-collector-cert\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.169224 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjprf\" (UniqueName: \"kubernetes.io/projected/c62d6a28-def0-41d5-a931-3bb27183deef-kube-api-access-wjprf\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.169262 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c62d6a28-def0-41d5-a931-3bb27183deef-srv-cert\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.169329 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c62d6a28-def0-41d5-a931-3bb27183deef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.169366 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.169433 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-proxy-tls\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.169647 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fb25\" (UniqueName: \"kubernetes.io/projected/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-kube-api-access-9fb25\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.170276 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.170550 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.171312 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efe68d-5663-488a-92ca-c02a288b524f-config-volume\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.174214 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c62d6a28-def0-41d5-a931-3bb27183deef-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.190386 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.210002 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.229831 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.232721 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/15cafda9-1dc9-40b0-af56-024dfaf32754-srv-cert\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.252037 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.271210 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.289803 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.293087 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.319272 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.320998 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.329541 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.350765 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.362683 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f273bc2e-26d7-40fe-9150-fb582fb14618-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-f6dtf\" (UID: \"f273bc2e-26d7-40fe-9150-fb582fb14618\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.371099 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.390410 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.403398 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c2151912-a994-4992-915d-d737497335d7-apiservice-cert\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.403486 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c2151912-a994-4992-915d-d737497335d7-webhook-cert\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.410423 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.431392 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.445631 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.449753 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.472036 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.480219 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.490694 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.523539 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff6pg\" (UniqueName: \"kubernetes.io/projected/20e8f550-62f2-4d57-9f9f-62123ac8bc2e-kube-api-access-ff6pg\") pod \"apiserver-76f77b778f-x6svn\" (UID: \"20e8f550-62f2-4d57-9f9f-62123ac8bc2e\") " pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.545063 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shld6\" (UniqueName: \"kubernetes.io/projected/69730259-7614-4869-bc84-10e17d8d13df-kube-api-access-shld6\") pod \"console-f9d7485db-wqc7t\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.554568 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.569749 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.589979 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.603149 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/acbd17d0-d2e2-4a68-991c-3553eb69f8ea-metrics-tls\") pod \"dns-operator-744455d44c-d7jk5\" (UID: \"acbd17d0-d2e2-4a68-991c-3553eb69f8ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.609743 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.630872 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.650091 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.670172 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.674609 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.690387 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.712267 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.729865 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.750738 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.769842 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.790299 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.809513 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.837824 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.837875 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-x6svn"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.849752 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.861014 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wqc7t"] Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.873274 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: W1010 09:13:17.877299 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69730259_7614_4869_bc84_10e17d8d13df.slice/crio-d7c3e8e876a4fd623fab152ba220421901928f6a9517bbef18f7ecf4013de2a2 WatchSource:0}: Error finding container d7c3e8e876a4fd623fab152ba220421901928f6a9517bbef18f7ecf4013de2a2: Status 404 returned error can't find the container with id d7c3e8e876a4fd623fab152ba220421901928f6a9517bbef18f7ecf4013de2a2 Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.889118 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.910380 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.928091 4669 request.go:700] Waited for 1.003451421s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dservice-ca-operator-dockercfg-rg9jl&limit=500&resourceVersion=0 Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.930015 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.949489 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.969734 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 10 09:13:17 crc kubenswrapper[4669]: E1010 09:13:17.973754 4669 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Oct 10 09:13:17 crc kubenswrapper[4669]: E1010 09:13:17.973859 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca podName:2a92e481-9a5c-4238-988f-f2d0b852ebca nodeName:}" failed. No retries permitted until 2025-10-10 09:13:18.473833385 +0000 UTC m=+141.489852127 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca") pod "route-controller-manager-6576b87f9c-l67bg" (UID: "2a92e481-9a5c-4238-988f-f2d0b852ebca") : failed to sync configmap cache: timed out waiting for the condition Oct 10 09:13:17 crc kubenswrapper[4669]: I1010 09:13:17.989807 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 10 09:13:17 crc kubenswrapper[4669]: E1010 09:13:17.992842 4669 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 10 09:13:17 crc kubenswrapper[4669]: E1010 09:13:17.992920 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert podName:2a92e481-9a5c-4238-988f-f2d0b852ebca nodeName:}" failed. No retries permitted until 2025-10-10 09:13:18.492896622 +0000 UTC m=+141.508915364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert") pod "route-controller-manager-6576b87f9c-l67bg" (UID: "2a92e481-9a5c-4238-988f-f2d0b852ebca") : failed to sync secret cache: timed out waiting for the condition Oct 10 09:13:17 crc kubenswrapper[4669]: E1010 09:13:17.997899 4669 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Oct 10 09:13:17 crc kubenswrapper[4669]: E1010 09:13:17.997950 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config podName:2a92e481-9a5c-4238-988f-f2d0b852ebca nodeName:}" failed. No retries permitted until 2025-10-10 09:13:18.497936594 +0000 UTC m=+141.513955336 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config") pod "route-controller-manager-6576b87f9c-l67bg" (UID: "2a92e481-9a5c-4238-988f-f2d0b852ebca") : failed to sync configmap cache: timed out waiting for the condition Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.009775 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.026125 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c62d6a28-def0-41d5-a931-3bb27183deef-srv-cert\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.030537 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.049741 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.069553 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.089938 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.109146 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.130323 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.149245 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 10 09:13:18 crc kubenswrapper[4669]: E1010 09:13:18.170215 4669 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Oct 10 09:13:18 crc kubenswrapper[4669]: E1010 09:13:18.170332 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-proxy-tls podName:8f9070f5-5a97-4cfd-9eea-9c5c011ea064 nodeName:}" failed. No retries permitted until 2025-10-10 09:13:18.670307683 +0000 UTC m=+141.686326435 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-proxy-tls") pod "machine-config-controller-84d6567774-stkm9" (UID: "8f9070f5-5a97-4cfd-9eea-9c5c011ea064") : failed to sync secret cache: timed out waiting for the condition Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.170410 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.189845 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.210421 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.230443 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.251070 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.271521 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.289982 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.329537 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf8wq\" (UniqueName: \"kubernetes.io/projected/eb510d78-5e60-48d3-8deb-3e7386845e0a-kube-api-access-gf8wq\") pod \"openshift-apiserver-operator-796bbdcf4f-wzpzn\" (UID: \"eb510d78-5e60-48d3-8deb-3e7386845e0a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.335042 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.369831 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxz7t\" (UniqueName: \"kubernetes.io/projected/55a8f8c1-7a6c-474f-b422-60d886882a59-kube-api-access-xxz7t\") pod \"openshift-config-operator-7777fb866f-kkf8r\" (UID: \"55a8f8c1-7a6c-474f-b422-60d886882a59\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.370399 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.390045 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.392974 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.410882 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.429890 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.463066 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.471236 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.476467 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.489635 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.517270 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9mv2\" (UniqueName: \"kubernetes.io/projected/6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7-kube-api-access-d9mv2\") pod \"service-ca-9c57cc56f-cr5vp\" (UID: \"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7\") " pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.525514 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq9qk\" (UniqueName: \"kubernetes.io/projected/0342148d-ffe9-4d1d-a629-03108f4c2cda-kube-api-access-kq9qk\") pod \"controller-manager-879f6c89f-8q8w9\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.549680 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm76k\" (UniqueName: \"kubernetes.io/projected/d3df4b15-909b-4a5d-af3a-af282d2ee197-kube-api-access-zm76k\") pod \"machine-config-operator-74547568cd-76vb5\" (UID: \"d3df4b15-909b-4a5d-af3a-af282d2ee197\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.574000 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.584821 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpxfm\" (UniqueName: \"kubernetes.io/projected/e57d53b2-5487-4820-aa61-9070eb866eae-kube-api-access-xpxfm\") pod \"machine-api-operator-5694c8668f-9lpfb\" (UID: \"e57d53b2-5487-4820-aa61-9070eb866eae\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.588922 4669 generic.go:334] "Generic (PLEG): container finished" podID="20e8f550-62f2-4d57-9f9f-62123ac8bc2e" containerID="d205c41ce624526f609794e2cfeb70e3ef7952ab4b3074fe77d68d7c747270c8" exitCode=0 Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.589018 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" event={"ID":"20e8f550-62f2-4d57-9f9f-62123ac8bc2e","Type":"ContainerDied","Data":"d205c41ce624526f609794e2cfeb70e3ef7952ab4b3074fe77d68d7c747270c8"} Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.589058 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" event={"ID":"20e8f550-62f2-4d57-9f9f-62123ac8bc2e","Type":"ContainerStarted","Data":"bb0a6bee5e3c3f8e11021a2ffef028d20785a41312f47b1b5fadf1868133fd21"} Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.590663 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.590750 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.603003 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wqc7t" event={"ID":"69730259-7614-4869-bc84-10e17d8d13df","Type":"ContainerStarted","Data":"e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b"} Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.603075 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wqc7t" event={"ID":"69730259-7614-4869-bc84-10e17d8d13df","Type":"ContainerStarted","Data":"d7c3e8e876a4fd623fab152ba220421901928f6a9517bbef18f7ecf4013de2a2"} Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.603924 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.618648 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.626077 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-584rk\" (UniqueName: \"kubernetes.io/projected/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-kube-api-access-584rk\") pod \"oauth-openshift-558db77b4-pb9vp\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.632726 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.645379 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjcth\" (UniqueName: \"kubernetes.io/projected/0aaee6b7-028b-447a-88ce-d70fc8c88bc6-kube-api-access-vjcth\") pod \"console-operator-58897d9998-nbx9n\" (UID: \"0aaee6b7-028b-447a-88ce-d70fc8c88bc6\") " pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.651140 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.675219 4669 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.691004 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.691793 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-proxy-tls\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.709735 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn"] Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.726111 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.730089 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsc44\" (UniqueName: \"kubernetes.io/projected/e5962f59-9740-46e8-9866-a57806721845-kube-api-access-wsc44\") pod \"apiserver-7bbb656c7d-zltk9\" (UID: \"e5962f59-9740-46e8-9866-a57806721845\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:18 crc kubenswrapper[4669]: W1010 09:13:18.737685 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb510d78_5e60_48d3_8deb_3e7386845e0a.slice/crio-66338506383c8ecf7e5a7dc42d430ff56a796f5fd6d5885af7cd10852ee22413 WatchSource:0}: Error finding container 66338506383c8ecf7e5a7dc42d430ff56a796f5fd6d5885af7cd10852ee22413: Status 404 returned error can't find the container with id 66338506383c8ecf7e5a7dc42d430ff56a796f5fd6d5885af7cd10852ee22413 Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.750005 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6clmf\" (UniqueName: \"kubernetes.io/projected/c6494b37-1d0e-4456-bfde-a0c48dcd6e1d-kube-api-access-6clmf\") pod \"machine-approver-56656f9798-klfh8\" (UID: \"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.764259 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq79w\" (UniqueName: \"kubernetes.io/projected/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-kube-api-access-nq79w\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.785777 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnk4t\" (UniqueName: \"kubernetes.io/projected/182c02d0-fdb0-4a01-bb6f-efdc3bbcb550-kube-api-access-qnk4t\") pod \"multus-admission-controller-857f4d67dd-rcqcc\" (UID: \"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.787032 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r"] Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.791227 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.794912 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.797519 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-proxy-tls\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.810335 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.820661 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.853417 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4j5h\" (UniqueName: \"kubernetes.io/projected/99fcf6de-fdb5-4759-9b09-01e59eca2655-kube-api-access-k4j5h\") pod \"openshift-controller-manager-operator-756b6f6bc6-khd8l\" (UID: \"99fcf6de-fdb5-4759-9b09-01e59eca2655\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.865668 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5"] Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.866040 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.869358 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmcpz\" (UniqueName: \"kubernetes.io/projected/52821f37-1b46-4913-b0ac-c00e0a899193-kube-api-access-cmcpz\") pod \"cluster-samples-operator-665b6dd947-j89vb\" (UID: \"52821f37-1b46-4913-b0ac-c00e0a899193\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.870340 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.896756 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.900580 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjds2\" (UniqueName: \"kubernetes.io/projected/521f5278-6996-4b67-a6ab-d46bd1a35acb-kube-api-access-pjds2\") pod \"migrator-59844c95c7-lvq4s\" (UID: \"521f5278-6996-4b67-a6ab-d46bd1a35acb\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.908088 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0e98983e-f7ef-4a56-a608-3715bc9f2ae7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-w5qk7\" (UID: \"0e98983e-f7ef-4a56-a608-3715bc9f2ae7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.908301 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.931463 4669 request.go:700] Waited for 1.887855973s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/default/token Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.931534 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.937563 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.937918 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8q8w9"] Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.942809 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk2dk\" (UniqueName: \"kubernetes.io/projected/24dc9ac3-3322-40f6-a99b-433280694265-kube-api-access-zk2dk\") pod \"authentication-operator-69f744f599-4mpz6\" (UID: \"24dc9ac3-3322-40f6-a99b-433280694265\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.965021 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" Oct 10 09:13:18 crc kubenswrapper[4669]: I1010 09:13:18.965704 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls7nt\" (UniqueName: \"kubernetes.io/projected/3f266776-152e-4b8c-b4e4-98a578b5f238-kube-api-access-ls7nt\") pod \"downloads-7954f5f757-ngvr9\" (UID: \"3f266776-152e-4b8c-b4e4-98a578b5f238\") " pod="openshift-console/downloads-7954f5f757-ngvr9" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.016244 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jffg9\" (UniqueName: \"kubernetes.io/projected/98efe68d-5663-488a-92ca-c02a288b524f-kube-api-access-jffg9\") pod \"collect-profiles-29334780-xjz79\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.018662 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m2ql\" (UniqueName: \"kubernetes.io/projected/89852882-ff51-4b3c-9e60-8a862b7d18cc-kube-api-access-7m2ql\") pod \"package-server-manager-789f6589d5-bngrg\" (UID: \"89852882-ff51-4b3c-9e60-8a862b7d18cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.039463 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ngvr9" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.048229 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrkw9\" (UniqueName: \"kubernetes.io/projected/acbd17d0-d2e2-4a68-991c-3553eb69f8ea-kube-api-access-wrkw9\") pod \"dns-operator-744455d44c-d7jk5\" (UID: \"acbd17d0-d2e2-4a68-991c-3553eb69f8ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.055746 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-cr5vp"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.057865 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkn84\" (UniqueName: \"kubernetes.io/projected/f273bc2e-26d7-40fe-9150-fb582fb14618-kube-api-access-jkn84\") pod \"control-plane-machine-set-operator-78cbb6b69f-f6dtf\" (UID: \"f273bc2e-26d7-40fe-9150-fb582fb14618\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.064315 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9lpfb"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.072509 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ct8g\" (UniqueName: \"kubernetes.io/projected/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-kube-api-access-7ct8g\") pod \"marketplace-operator-79b997595-nvfcj\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:19 crc kubenswrapper[4669]: W1010 09:13:19.077765 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6aaf22e6_5ebb_42b7_b580_c2f1745f1bd7.slice/crio-ac5a4c218c6d12a4e20c6a9a043eb262a23549578c96cbcdf9ec49cc5818b0f2 WatchSource:0}: Error finding container ac5a4c218c6d12a4e20c6a9a043eb262a23549578c96cbcdf9ec49cc5818b0f2: Status 404 returned error can't find the container with id ac5a4c218c6d12a4e20c6a9a043eb262a23549578c96cbcdf9ec49cc5818b0f2 Oct 10 09:13:19 crc kubenswrapper[4669]: W1010 09:13:19.091527 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode57d53b2_5487_4820_aa61_9070eb866eae.slice/crio-7e97061dc2bba690a704bf72488c0c9bf0e34eccde529dd9fbad66831369ed10 WatchSource:0}: Error finding container 7e97061dc2bba690a704bf72488c0c9bf0e34eccde529dd9fbad66831369ed10: Status 404 returned error can't find the container with id 7e97061dc2bba690a704bf72488c0c9bf0e34eccde529dd9fbad66831369ed10 Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.101187 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q62c6\" (UniqueName: \"kubernetes.io/projected/6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1-kube-api-access-q62c6\") pod \"kube-storage-version-migrator-operator-b67b599dd-xk9nh\" (UID: \"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.105953 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8v7p\" (UniqueName: \"kubernetes.io/projected/c2151912-a994-4992-915d-d737497335d7-kube-api-access-f8v7p\") pod \"packageserver-d55dfcdfc-kdgw8\" (UID: \"c2151912-a994-4992-915d-d737497335d7\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.126270 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqdcx\" (UniqueName: \"kubernetes.io/projected/15cafda9-1dc9-40b0-af56-024dfaf32754-kube-api-access-wqdcx\") pod \"catalog-operator-68c6474976-2jf4j\" (UID: \"15cafda9-1dc9-40b0-af56-024dfaf32754\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.136931 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.149993 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.177957 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.179309 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.193022 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.196089 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nbx9n"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.212011 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.238859 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.241522 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.246951 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.247091 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.253662 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.260853 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.261867 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.271497 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.271927 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.278091 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.285409 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.291096 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.292654 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.310281 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.353294 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjprf\" (UniqueName: \"kubernetes.io/projected/c62d6a28-def0-41d5-a931-3bb27183deef-kube-api-access-wjprf\") pod \"olm-operator-6b444d44fb-rwgqq\" (UID: \"c62d6a28-def0-41d5-a931-3bb27183deef\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.375692 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.381443 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fb25\" (UniqueName: \"kubernetes.io/projected/8f9070f5-5a97-4cfd-9eea-9c5c011ea064-kube-api-access-9fb25\") pod \"machine-config-controller-84d6567774-stkm9\" (UID: \"8f9070f5-5a97-4cfd-9eea-9c5c011ea064\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.391605 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.421190 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.424225 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.446777 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.453889 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.455755 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.469718 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.485019 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggp22\" (UniqueName: \"kubernetes.io/projected/2a92e481-9a5c-4238-988f-f2d0b852ebca-kube-api-access-ggp22\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.490398 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.492932 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config\") pod \"route-controller-manager-6576b87f9c-l67bg\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.511791 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.551861 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613207 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfb7109a-2f1c-49c9-960b-0be9a7028b35-ca-trust-extracted\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613237 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-config\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613305 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbmhr\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-kube-api-access-zbmhr\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613332 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-trusted-ca\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613397 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4d82f5-1059-46f6-a639-664fd1e47963-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613416 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-default-certificate\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613463 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-registration-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613478 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5783db5c-e7ff-4423-9188-f75344b4a111-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613496 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgfxf\" (UniqueName: \"kubernetes.io/projected/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-kube-api-access-wgfxf\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613535 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4d82f5-1059-46f6-a639-664fd1e47963-config\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613550 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b4d82f5-1059-46f6-a639-664fd1e47963-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.613565 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5783db5c-e7ff-4423-9188-f75344b4a111-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618309 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-bound-sa-token\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618573 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-service-ca\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618621 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-serving-cert\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618675 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfb7109a-2f1c-49c9-960b-0be9a7028b35-installation-pull-secrets\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618691 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-config\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618742 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3c7621d-224f-4ada-954c-5f9f051a6132-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618759 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-csi-data-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618775 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hn6g\" (UniqueName: \"kubernetes.io/projected/201e2060-b78b-4f0b-a9a7-6855bd9158eb-kube-api-access-9hn6g\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618789 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c7621d-224f-4ada-954c-5f9f051a6132-config\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618820 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlhz8\" (UniqueName: \"kubernetes.io/projected/2f33e625-27c7-4380-be4c-8468bde0feb7-kube-api-access-rlhz8\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.618908 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-client\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619039 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-tls\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619108 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5783db5c-e7ff-4423-9188-f75344b4a111-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619127 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-socket-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619174 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-stats-auth\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619213 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-ca\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619268 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f33e625-27c7-4380-be4c-8468bde0feb7-trusted-ca\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619312 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-metrics-certs\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619344 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-certificates\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619393 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c7621d-224f-4ada-954c-5f9f051a6132-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619424 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m77n6\" (UniqueName: \"kubernetes.io/projected/e505e6fd-e00f-40a8-a575-6102ec52059d-kube-api-access-m77n6\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619479 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-serving-cert\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619515 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-plugins-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619633 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2f33e625-27c7-4380-be4c-8468bde0feb7-metrics-tls\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.619670 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2g5t\" (UniqueName: \"kubernetes.io/projected/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-kube-api-access-j2g5t\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.620163 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.620184 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-mountpoint-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.620315 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f33e625-27c7-4380-be4c-8468bde0feb7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.626188 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e505e6fd-e00f-40a8-a575-6102ec52059d-service-ca-bundle\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: E1010 09:13:19.633811 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.131841753 +0000 UTC m=+143.147860495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.650018 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.711571 4669 generic.go:334] "Generic (PLEG): container finished" podID="55a8f8c1-7a6c-474f-b422-60d886882a59" containerID="f830751717ef6ed459de22e93a003d6ef2b34559b094c2b7b403a25457b5c8f0" exitCode=0 Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.711678 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" event={"ID":"55a8f8c1-7a6c-474f-b422-60d886882a59","Type":"ContainerDied","Data":"f830751717ef6ed459de22e93a003d6ef2b34559b094c2b7b403a25457b5c8f0"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.711703 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" event={"ID":"55a8f8c1-7a6c-474f-b422-60d886882a59","Type":"ContainerStarted","Data":"d57584a286bab3819f8ec31c410b8b232551e96f14c8cf986f4dd10b4d3f206d"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727258 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727634 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-certs\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727728 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-registration-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727766 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5783db5c-e7ff-4423-9188-f75344b4a111-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727808 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgfxf\" (UniqueName: \"kubernetes.io/projected/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-kube-api-access-wgfxf\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727835 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cklp\" (UniqueName: \"kubernetes.io/projected/57749860-6c06-44f2-bbd9-382cf3bbd3c8-kube-api-access-4cklp\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727879 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4d82f5-1059-46f6-a639-664fd1e47963-config\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727903 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b4d82f5-1059-46f6-a639-664fd1e47963-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727929 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5783db5c-e7ff-4423-9188-f75344b4a111-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.727973 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-bound-sa-token\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728039 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-service-ca\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728066 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-serving-cert\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728095 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfb7109a-2f1c-49c9-960b-0be9a7028b35-installation-pull-secrets\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728133 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-config\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728161 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3c7621d-224f-4ada-954c-5f9f051a6132-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728183 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-csi-data-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728204 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hn6g\" (UniqueName: \"kubernetes.io/projected/201e2060-b78b-4f0b-a9a7-6855bd9158eb-kube-api-access-9hn6g\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728227 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c7621d-224f-4ada-954c-5f9f051a6132-config\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728252 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlhz8\" (UniqueName: \"kubernetes.io/projected/2f33e625-27c7-4380-be4c-8468bde0feb7-kube-api-access-rlhz8\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728280 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-client\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728311 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-tls\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728337 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ec9ad95-b8ea-459b-a151-1601edf706ea-cert\") pod \"ingress-canary-llrlv\" (UID: \"6ec9ad95-b8ea-459b-a151-1601edf706ea\") " pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728361 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5783db5c-e7ff-4423-9188-f75344b4a111-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728380 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-socket-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728404 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnc7r\" (UniqueName: \"kubernetes.io/projected/6ec9ad95-b8ea-459b-a151-1601edf706ea-kube-api-access-bnc7r\") pod \"ingress-canary-llrlv\" (UID: \"6ec9ad95-b8ea-459b-a151-1601edf706ea\") " pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728429 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-stats-auth\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728452 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-ca\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrlns\" (UniqueName: \"kubernetes.io/projected/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-kube-api-access-mrlns\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728495 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f33e625-27c7-4380-be4c-8468bde0feb7-trusted-ca\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728514 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-node-bootstrap-token\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728541 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-metrics-certs\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728572 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-certificates\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.728611 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c7621d-224f-4ada-954c-5f9f051a6132-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: E1010 09:13:19.729921 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.229890761 +0000 UTC m=+143.245909683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.730544 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-registration-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.731207 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-socket-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.731313 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-csi-data-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.732213 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4d82f5-1059-46f6-a639-664fd1e47963-config\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733077 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m77n6\" (UniqueName: \"kubernetes.io/projected/e505e6fd-e00f-40a8-a575-6102ec52059d-kube-api-access-m77n6\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733114 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-serving-cert\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733138 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-plugins-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733174 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/57749860-6c06-44f2-bbd9-382cf3bbd3c8-metrics-tls\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733222 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2f33e625-27c7-4380-be4c-8468bde0feb7-metrics-tls\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733245 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2g5t\" (UniqueName: \"kubernetes.io/projected/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-kube-api-access-j2g5t\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733272 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733289 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-mountpoint-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733335 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f33e625-27c7-4380-be4c-8468bde0feb7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733355 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57749860-6c06-44f2-bbd9-382cf3bbd3c8-config-volume\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733388 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e505e6fd-e00f-40a8-a575-6102ec52059d-service-ca-bundle\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733407 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfb7109a-2f1c-49c9-960b-0be9a7028b35-ca-trust-extracted\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733426 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-config\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733488 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbmhr\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-kube-api-access-zbmhr\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.733546 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-trusted-ca\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.735723 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4d82f5-1059-46f6-a639-664fd1e47963-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.735759 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-default-certificate\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.738833 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5783db5c-e7ff-4423-9188-f75344b4a111-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.741566 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-service-ca\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.748561 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e505e6fd-e00f-40a8-a575-6102ec52059d-service-ca-bundle\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.748944 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfb7109a-2f1c-49c9-960b-0be9a7028b35-ca-trust-extracted\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.749356 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-config\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.754916 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5783db5c-e7ff-4423-9188-f75344b4a111-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: E1010 09:13:19.761248 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.261228574 +0000 UTC m=+143.277247316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.763093 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-serving-cert\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.763519 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3c7621d-224f-4ada-954c-5f9f051a6132-config\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.766034 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-default-certificate\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.766308 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfb7109a-2f1c-49c9-960b-0be9a7028b35-installation-pull-secrets\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.766660 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-config\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.769445 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4d82f5-1059-46f6-a639-664fd1e47963-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.769931 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2f33e625-27c7-4380-be4c-8468bde0feb7-metrics-tls\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.771703 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-mountpoint-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.773021 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-ca\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.776892 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-trusted-ca\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.778244 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3c7621d-224f-4ada-954c-5f9f051a6132-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.778628 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-stats-auth\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.779067 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f33e625-27c7-4380-be4c-8468bde0feb7-trusted-ca\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.779555 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-tls\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.785989 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-etcd-client\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.786264 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" event={"ID":"0aaee6b7-028b-447a-88ce-d70fc8c88bc6","Type":"ContainerStarted","Data":"48d76420c4dedad098f541d2b6a47da587bce01ff6fd1ae88b47a0789c13ddb5"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.789888 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-certificates\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.792208 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/201e2060-b78b-4f0b-a9a7-6855bd9158eb-plugins-dir\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.798580 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-serving-cert\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.800033 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3c7621d-224f-4ada-954c-5f9f051a6132-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r6blz\" (UID: \"d3c7621d-224f-4ada-954c-5f9f051a6132\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.830654 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e505e6fd-e00f-40a8-a575-6102ec52059d-metrics-certs\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.831462 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hn6g\" (UniqueName: \"kubernetes.io/projected/201e2060-b78b-4f0b-a9a7-6855bd9158eb-kube-api-access-9hn6g\") pod \"csi-hostpathplugin-mfjcd\" (UID: \"201e2060-b78b-4f0b-a9a7-6855bd9158eb\") " pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.833352 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5783db5c-e7ff-4423-9188-f75344b4a111-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-8226f\" (UID: \"5783db5c-e7ff-4423-9188-f75344b4a111\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.841332 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:19 crc kubenswrapper[4669]: E1010 09:13:19.841530 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.341494777 +0000 UTC m=+143.357513519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.841598 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/57749860-6c06-44f2-bbd9-382cf3bbd3c8-metrics-tls\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.841655 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.841704 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57749860-6c06-44f2-bbd9-382cf3bbd3c8-config-volume\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.841762 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-certs\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.841798 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cklp\" (UniqueName: \"kubernetes.io/projected/57749860-6c06-44f2-bbd9-382cf3bbd3c8-kube-api-access-4cklp\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: E1010 09:13:19.842990 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.342974315 +0000 UTC m=+143.358993057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.848773 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ec9ad95-b8ea-459b-a151-1601edf706ea-cert\") pod \"ingress-canary-llrlv\" (UID: \"6ec9ad95-b8ea-459b-a151-1601edf706ea\") " pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.848822 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnc7r\" (UniqueName: \"kubernetes.io/projected/6ec9ad95-b8ea-459b-a151-1601edf706ea-kube-api-access-bnc7r\") pod \"ingress-canary-llrlv\" (UID: \"6ec9ad95-b8ea-459b-a151-1601edf706ea\") " pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.848866 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrlns\" (UniqueName: \"kubernetes.io/projected/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-kube-api-access-mrlns\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.848892 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-node-bootstrap-token\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.855286 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57749860-6c06-44f2-bbd9-382cf3bbd3c8-config-volume\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.862453 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" event={"ID":"d3df4b15-909b-4a5d-af3a-af282d2ee197","Type":"ContainerStarted","Data":"058ec398d3328a88f6dd0e9a339cac4f6c4778e4ca172ea95c90d74e5ff6819f"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.862490 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb9vp"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.862506 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nvfcj"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.862516 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.862527 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" event={"ID":"d3df4b15-909b-4a5d-af3a-af282d2ee197","Type":"ContainerStarted","Data":"e026f078541297944a46bb2ab0eca35486defa526b0cf9f6c77d449cdb51506b"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.862538 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rcqcc"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.862551 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ngvr9"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.868947 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.871713 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgfxf\" (UniqueName: \"kubernetes.io/projected/033ee508-b7b5-47cf-984e-6d1c2e41b8ef-kube-api-access-wgfxf\") pod \"service-ca-operator-777779d784-4xs8s\" (UID: \"033ee508-b7b5-47cf-984e-6d1c2e41b8ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.874929 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-bound-sa-token\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.884919 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ec9ad95-b8ea-459b-a151-1601edf706ea-cert\") pod \"ingress-canary-llrlv\" (UID: \"6ec9ad95-b8ea-459b-a151-1601edf706ea\") " pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.887279 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b4d82f5-1059-46f6-a639-664fd1e47963-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-grdvw\" (UID: \"7b4d82f5-1059-46f6-a639-664fd1e47963\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.890099 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/57749860-6c06-44f2-bbd9-382cf3bbd3c8-metrics-tls\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.902482 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-certs\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.903664 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-node-bootstrap-token\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.905195 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4mpz6"] Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.915726 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.921145 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlhz8\" (UniqueName: \"kubernetes.io/projected/2f33e625-27c7-4380-be4c-8468bde0feb7-kube-api-access-rlhz8\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.931220 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" event={"ID":"e5962f59-9740-46e8-9866-a57806721845","Type":"ContainerStarted","Data":"78553811b2d009b2850b9656c2b09a879da52acf0105a18383f099e6087d3507"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.933338 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7"] Oct 10 09:13:19 crc kubenswrapper[4669]: W1010 09:13:19.937924 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f266776_152e_4b8c_b4e4_98a578b5f238.slice/crio-b0f454c5cf0ce59d5d9891a7984b82a8610fc4bebf3680624b400f3aee30fe5f WatchSource:0}: Error finding container b0f454c5cf0ce59d5d9891a7984b82a8610fc4bebf3680624b400f3aee30fe5f: Status 404 returned error can't find the container with id b0f454c5cf0ce59d5d9891a7984b82a8610fc4bebf3680624b400f3aee30fe5f Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.941047 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbmhr\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-kube-api-access-zbmhr\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.942137 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" event={"ID":"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d","Type":"ContainerStarted","Data":"8135c5a45339f969fcb079c70a750d08d844e2f763eed3be459b35191529054b"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.942168 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" event={"ID":"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d","Type":"ContainerStarted","Data":"d330ed063f7f9ad1ca8f8b1e8478544c5c91c6f8b929809983b868e9b6057d4f"} Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.945862 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2g5t\" (UniqueName: \"kubernetes.io/projected/caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a-kube-api-access-j2g5t\") pod \"etcd-operator-b45778765-klgpb\" (UID: \"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.952118 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.955234 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:19 crc kubenswrapper[4669]: E1010 09:13:19.956217 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.456194482 +0000 UTC m=+143.472213224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.960130 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f33e625-27c7-4380-be4c-8468bde0feb7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-xvhtj\" (UID: \"2f33e625-27c7-4380-be4c-8468bde0feb7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.963832 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.979190 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" Oct 10 09:13:19 crc kubenswrapper[4669]: I1010 09:13:19.980617 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.002650 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m77n6\" (UniqueName: \"kubernetes.io/projected/e505e6fd-e00f-40a8-a575-6102ec52059d-kube-api-access-m77n6\") pod \"router-default-5444994796-zqhsj\" (UID: \"e505e6fd-e00f-40a8-a575-6102ec52059d\") " pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.005315 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cklp\" (UniqueName: \"kubernetes.io/projected/57749860-6c06-44f2-bbd9-382cf3bbd3c8-kube-api-access-4cklp\") pod \"dns-default-rnpbk\" (UID: \"57749860-6c06-44f2-bbd9-382cf3bbd3c8\") " pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.016902 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.030330 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnc7r\" (UniqueName: \"kubernetes.io/projected/6ec9ad95-b8ea-459b-a151-1601edf706ea-kube-api-access-bnc7r\") pod \"ingress-canary-llrlv\" (UID: \"6ec9ad95-b8ea-459b-a151-1601edf706ea\") " pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.035846 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.045408 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" event={"ID":"eb510d78-5e60-48d3-8deb-3e7386845e0a","Type":"ContainerStarted","Data":"80e1100245cdf09494689d543b3b0a4ad6ee01c283b51974d42bf4d37de96bc4"} Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.045449 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" event={"ID":"eb510d78-5e60-48d3-8deb-3e7386845e0a","Type":"ContainerStarted","Data":"66338506383c8ecf7e5a7dc42d430ff56a796f5fd6d5885af7cd10852ee22413"} Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.047781 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-llrlv" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.058060 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.058611 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.558580001 +0000 UTC m=+143.574598743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.071288 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrlns\" (UniqueName: \"kubernetes.io/projected/d0b0ecc3-1f98-4345-8095-2a69aedeeac5-kube-api-access-mrlns\") pod \"machine-config-server-hsgp7\" (UID: \"d0b0ecc3-1f98-4345-8095-2a69aedeeac5\") " pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:20 crc kubenswrapper[4669]: W1010 09:13:20.096711 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e98983e_f7ef_4a56_a608_3715bc9f2ae7.slice/crio-2bd4a544c73f850ee26fd522cfd7d6d2bf34cfd63be54d52a95b463aa777941a WatchSource:0}: Error finding container 2bd4a544c73f850ee26fd522cfd7d6d2bf34cfd63be54d52a95b463aa777941a: Status 404 returned error can't find the container with id 2bd4a544c73f850ee26fd522cfd7d6d2bf34cfd63be54d52a95b463aa777941a Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.129641 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.160031 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.164578 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.664539214 +0000 UTC m=+143.680557956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.177414 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" event={"ID":"20e8f550-62f2-4d57-9f9f-62123ac8bc2e","Type":"ContainerStarted","Data":"aeb3b45eda05df9aaa9f15743800afbf44fecf4f6e99aa94f19082dc55d7a65a"} Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.202028 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.208679 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.220874 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" event={"ID":"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7","Type":"ContainerStarted","Data":"ac5a4c218c6d12a4e20c6a9a043eb262a23549578c96cbcdf9ec49cc5818b0f2"} Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.226531 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" event={"ID":"e57d53b2-5487-4820-aa61-9070eb866eae","Type":"ContainerStarted","Data":"7e97061dc2bba690a704bf72488c0c9bf0e34eccde529dd9fbad66831369ed10"} Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.229719 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.231947 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" event={"ID":"0342148d-ffe9-4d1d-a629-03108f4c2cda","Type":"ContainerStarted","Data":"f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739"} Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.232094 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" event={"ID":"0342148d-ffe9-4d1d-a629-03108f4c2cda","Type":"ContainerStarted","Data":"13102c2875d2040863f0c3bb8410ca94fb7823f14721508cf3794edb55e959e7"} Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.232737 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.250788 4669 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8q8w9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.251131 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" podUID="0342148d-ffe9-4d1d-a629-03108f4c2cda" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.262146 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.271180 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.771144168 +0000 UTC m=+143.787162910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.348482 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hsgp7" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.365952 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.366073 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.866037654 +0000 UTC m=+143.882056396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.366482 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.367353 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.867344966 +0000 UTC m=+143.883363708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.468759 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.469098 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.969080733 +0000 UTC m=+143.985099475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.469503 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.469881 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:20.969866549 +0000 UTC m=+143.985885291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.477275 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.489783 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.507658 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.530030 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-d7jk5"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.570322 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.570798 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.070778629 +0000 UTC m=+144.086797371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.605139 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.621365 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.656550 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.674313 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.674671 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.174659886 +0000 UTC m=+144.190678628 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.713476 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.775258 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.775739 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.275720781 +0000 UTC m=+144.291739523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.801507 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq"] Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.876381 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.877047 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.377032873 +0000 UTC m=+144.393051615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.902061 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wqc7t" podStartSLOduration=121.902047551 podStartE2EDuration="2m1.902047551s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:20.900530793 +0000 UTC m=+143.916549535" watchObservedRunningTime="2025-10-10 09:13:20.902047551 +0000 UTC m=+143.918066293" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.954454 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wzpzn" podStartSLOduration=121.954437485 podStartE2EDuration="2m1.954437485s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:20.953469663 +0000 UTC m=+143.969488405" watchObservedRunningTime="2025-10-10 09:13:20.954437485 +0000 UTC m=+143.970456217" Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.979155 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.980549 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.480349171 +0000 UTC m=+144.496367913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:20 crc kubenswrapper[4669]: I1010 09:13:20.980729 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:20 crc kubenswrapper[4669]: E1010 09:13:20.981436 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.481427486 +0000 UTC m=+144.497446228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.082347 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.082680 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.582663207 +0000 UTC m=+144.598681949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: W1010 09:13:21.158713 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc62d6a28_def0_41d5_a931_3bb27183deef.slice/crio-19bf896a2b8ccef305a702b85b4b2a19b69a3530db442f0e790ca2ed4659a158 WatchSource:0}: Error finding container 19bf896a2b8ccef305a702b85b4b2a19b69a3530db442f0e790ca2ed4659a158: Status 404 returned error can't find the container with id 19bf896a2b8ccef305a702b85b4b2a19b69a3530db442f0e790ca2ed4659a158 Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.184428 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.184853 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.684832858 +0000 UTC m=+144.700851600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.270063 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" event={"ID":"033ee508-b7b5-47cf-984e-6d1c2e41b8ef","Type":"ContainerStarted","Data":"eec3b7df91d956bb0972e1f5a6abfe258102ab089a7eb5468d872c4e5ac291a1"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.274254 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" event={"ID":"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4","Type":"ContainerStarted","Data":"84bdc7a3323e421ad75902dae8354f3b9df2bbc4e2d342daa63dc0e93b1c1678"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.276913 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" event={"ID":"55a8f8c1-7a6c-474f-b422-60d886882a59","Type":"ContainerStarted","Data":"b501887ec35e4cedaf9fbc24ae9511592e9bc4130ff37d2ef8021e41723293c1"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.276947 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.286396 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.286726 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.786682379 +0000 UTC m=+144.802701121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.287133 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.287529 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.787505036 +0000 UTC m=+144.803523778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.295886 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" event={"ID":"99fcf6de-fdb5-4759-9b09-01e59eca2655","Type":"ContainerStarted","Data":"944ccc66f8899583fbdbc55d021fdff9f223c6fe6bf008692d1260303bced8b5"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.295942 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" event={"ID":"99fcf6de-fdb5-4759-9b09-01e59eca2655","Type":"ContainerStarted","Data":"fdef8e40d139188067608792a83d47440f955c8338266980e33be46a49c406b8"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.305716 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" event={"ID":"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550","Type":"ContainerStarted","Data":"3823fc8a03f595d6a4f2b5f575730528a50fae6b3127fbd507f88297cb87cdb7"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.314301 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" podStartSLOduration=122.31428231 podStartE2EDuration="2m2.31428231s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:21.290004136 +0000 UTC m=+144.306022878" watchObservedRunningTime="2025-10-10 09:13:21.31428231 +0000 UTC m=+144.330301052" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.319388 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rnpbk"] Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.351441 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" event={"ID":"e57d53b2-5487-4820-aa61-9070eb866eae","Type":"ContainerStarted","Data":"f6c50e263c3aaccd7efb3e58243fc57022320afd36c6c58cadcf85f2b2ef5896"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.366518 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" event={"ID":"52821f37-1b46-4913-b0ac-c00e0a899193","Type":"ContainerStarted","Data":"24a1387710d35d7a3e871a0719a14b17ef8e52707580465d4c71e60939e888c9"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.377963 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" event={"ID":"89852882-ff51-4b3c-9e60-8a862b7d18cc","Type":"ContainerStarted","Data":"90bab1c5694bd6b7f0383057489a395a4221fae33bec0246879d3eabe32385a1"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.378016 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" event={"ID":"89852882-ff51-4b3c-9e60-8a862b7d18cc","Type":"ContainerStarted","Data":"4b0d8a14007175775e08c99bfb3373eb36f7180570ba8d4913ba2d8d615bc349"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.387247 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" event={"ID":"01b28f7e-7133-451a-b0d5-d1c0e13b3b40","Type":"ContainerStarted","Data":"18a0d221e8a00be5e90902c7004cc94602474bd7816f6a2bc0293f668cc84699"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.387282 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" event={"ID":"01b28f7e-7133-451a-b0d5-d1c0e13b3b40","Type":"ContainerStarted","Data":"266e561eed73bad04b5945bb98d9d357441048f527fc1e6188af0b9e0b3d0056"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.388314 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.391327 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.393975 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.893942355 +0000 UTC m=+144.909961107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.394867 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" event={"ID":"c62d6a28-def0-41d5-a931-3bb27183deef","Type":"ContainerStarted","Data":"19bf896a2b8ccef305a702b85b4b2a19b69a3530db442f0e790ca2ed4659a158"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.399888 4669 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nvfcj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.399954 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" podUID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.434249 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" event={"ID":"d3df4b15-909b-4a5d-af3a-af282d2ee197","Type":"ContainerStarted","Data":"34129ba4b883c952c3facafb695f07604d537d5e495407c2b314bf3a917ddb5e"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.448732 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" event={"ID":"acbd17d0-d2e2-4a68-991c-3553eb69f8ea","Type":"ContainerStarted","Data":"476820777a8b8906779aeedee96a9b7b0d33ad1f185f56a10f55150a82f05c6b"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.450518 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" event={"ID":"0aaee6b7-028b-447a-88ce-d70fc8c88bc6","Type":"ContainerStarted","Data":"5997d50dab68c79313fbfaa259bf37d6125e3c78d8d5abf060a60cb7a273d91f"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.451770 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.463526 4669 patch_prober.go:28] interesting pod/console-operator-58897d9998-nbx9n container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.463599 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" podUID="0aaee6b7-028b-447a-88ce-d70fc8c88bc6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.467537 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" event={"ID":"f273bc2e-26d7-40fe-9150-fb582fb14618","Type":"ContainerStarted","Data":"cc42e84b86dd94dbdf2399e10ba9e88cc867d42057ad30421cae4c0927cd0be8"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.469377 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zqhsj" event={"ID":"e505e6fd-e00f-40a8-a575-6102ec52059d","Type":"ContainerStarted","Data":"2c66ccaff106363b64cdeea11496771da3ce41bc8d447f40bafda4fbb3b4418e"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.492732 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.493214 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:21.993202081 +0000 UTC m=+145.009220823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.550000 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" event={"ID":"20e8f550-62f2-4d57-9f9f-62123ac8bc2e","Type":"ContainerStarted","Data":"ffa9e26c40c05e870979dcc4dc143e1182757ee7d52e58add862b4f613b9c348"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.558026 4669 generic.go:334] "Generic (PLEG): container finished" podID="e5962f59-9740-46e8-9866-a57806721845" containerID="20c32e76b8314c89d461c3f22b0baac23ef2827e8964b9759b0dc29e9698396e" exitCode=0 Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.558110 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" event={"ID":"e5962f59-9740-46e8-9866-a57806721845","Type":"ContainerDied","Data":"20c32e76b8314c89d461c3f22b0baac23ef2827e8964b9759b0dc29e9698396e"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.594486 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.596849 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.096828699 +0000 UTC m=+145.112847441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.634483 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" event={"ID":"24dc9ac3-3322-40f6-a99b-433280694265","Type":"ContainerStarted","Data":"023afbe1f45e3ca36ba0b1531c3e26efd2a3eec2334af427771640becb7a9737"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.639713 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f"] Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.657368 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ngvr9" event={"ID":"3f266776-152e-4b8c-b4e4-98a578b5f238","Type":"ContainerStarted","Data":"b0f454c5cf0ce59d5d9891a7984b82a8610fc4bebf3680624b400f3aee30fe5f"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.722416 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" podStartSLOduration=122.722398156 podStartE2EDuration="2m2.722398156s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:21.721776716 +0000 UTC m=+144.737795458" watchObservedRunningTime="2025-10-10 09:13:21.722398156 +0000 UTC m=+144.738416898" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.725668 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.726728 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.226715366 +0000 UTC m=+145.242734108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.767629 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" podStartSLOduration=122.767608807 podStartE2EDuration="2m2.767608807s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:21.765048724 +0000 UTC m=+144.781067466" watchObservedRunningTime="2025-10-10 09:13:21.767608807 +0000 UTC m=+144.783627549" Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.779263 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz"] Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.793441 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" event={"ID":"c6494b37-1d0e-4456-bfde-a0c48dcd6e1d","Type":"ContainerStarted","Data":"288882a2a9f18bbeb83abaca5b04cb950af5ffdab13ebfd315befdf5d6d43d8a"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.824656 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" event={"ID":"15cafda9-1dc9-40b0-af56-024dfaf32754","Type":"ContainerStarted","Data":"99c20f1a6c6ab1100ca471ff8487e62e915c79ad4303d6a10bcdc1cdda0e4897"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.824978 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" event={"ID":"c2151912-a994-4992-915d-d737497335d7","Type":"ContainerStarted","Data":"b14f1b6c6ad92f6f96e39d1ff1dab44bd37874a54c67ae87b029d692dcd32188"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.826896 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" event={"ID":"2a92e481-9a5c-4238-988f-f2d0b852ebca","Type":"ContainerStarted","Data":"6f13ad89718ae3f7c0a8663fcfa00b19f084320ef2b57596a0462a857ac5faf0"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.838355 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.839800 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.339774258 +0000 UTC m=+145.355793010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.855956 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" event={"ID":"521f5278-6996-4b67-a6ab-d46bd1a35acb","Type":"ContainerStarted","Data":"aecc19f8a25515ec00f444298309a9796299c02aa84e5c16fc7cb3e772635513"} Oct 10 09:13:21 crc kubenswrapper[4669]: W1010 09:13:21.870141 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5783db5c_e7ff_4423_9188_f75344b4a111.slice/crio-55b5eb84ae66b19498ff632a81565b877d62d52f1e67b648eb77fdc4740ab461 WatchSource:0}: Error finding container 55b5eb84ae66b19498ff632a81565b877d62d52f1e67b648eb77fdc4740ab461: Status 404 returned error can't find the container with id 55b5eb84ae66b19498ff632a81565b877d62d52f1e67b648eb77fdc4740ab461 Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.914424 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" event={"ID":"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1","Type":"ContainerStarted","Data":"35ae23aff16e4aae8ee9155dc22f006bac18855bc5902b56aaae656465d6fc5a"} Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.944217 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:21 crc kubenswrapper[4669]: E1010 09:13:21.944845 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.444830523 +0000 UTC m=+145.460849265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:21 crc kubenswrapper[4669]: I1010 09:13:21.958698 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" event={"ID":"8f9070f5-5a97-4cfd-9eea-9c5c011ea064","Type":"ContainerStarted","Data":"3179e4ddfbc5cb331034fab1f83d2a90329de17eacc862ccf5c5f3a30dc82706"} Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.006074 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" event={"ID":"0e98983e-f7ef-4a56-a608-3715bc9f2ae7","Type":"ContainerStarted","Data":"2bd4a544c73f850ee26fd522cfd7d6d2bf34cfd63be54d52a95b463aa777941a"} Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.031448 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" podStartSLOduration=123.031431681 podStartE2EDuration="2m3.031431681s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:21.84819076 +0000 UTC m=+144.864209532" watchObservedRunningTime="2025-10-10 09:13:22.031431681 +0000 UTC m=+145.047450423" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.044974 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.045449 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.545429323 +0000 UTC m=+145.561448065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.059257 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" event={"ID":"98efe68d-5663-488a-92ca-c02a288b524f","Type":"ContainerStarted","Data":"c55ca4c5ebf485127a24b74d1c9420306b46f7d4ba625b42fe2eb58ec6ff26db"} Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.083375 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-klgpb"] Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.109836 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" event={"ID":"6aaf22e6-5ebb-42b7-b580-c2f1745f1bd7","Type":"ContainerStarted","Data":"70c4c8e24c7a6806c8ea3267a730ecc6da209d8d93126a1e9fa462f5344ea22b"} Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.133921 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.153225 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.153796 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.653779913 +0000 UTC m=+145.669798655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.220121 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mfjcd"] Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.223379 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-76vb5" podStartSLOduration=123.223353411 podStartE2EDuration="2m3.223353411s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:22.213243825 +0000 UTC m=+145.229262567" watchObservedRunningTime="2025-10-10 09:13:22.223353411 +0000 UTC m=+145.239372153" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.254552 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.265638 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.765605336 +0000 UTC m=+145.781624088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.267443 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.269972 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.769948767 +0000 UTC m=+145.785967709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.283301 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" podStartSLOduration=123.283279947 podStartE2EDuration="2m3.283279947s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:22.282898425 +0000 UTC m=+145.298917177" watchObservedRunningTime="2025-10-10 09:13:22.283279947 +0000 UTC m=+145.299298689" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.289624 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj"] Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.334930 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-khd8l" podStartSLOduration=123.334906516 podStartE2EDuration="2m3.334906516s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:22.332366444 +0000 UTC m=+145.348385186" watchObservedRunningTime="2025-10-10 09:13:22.334906516 +0000 UTC m=+145.350925258" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.403707 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.404187 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.904155583 +0000 UTC m=+145.920174325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.405100 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.405454 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:22.905438064 +0000 UTC m=+145.921456806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.418002 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-llrlv"] Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.458507 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-cr5vp" podStartSLOduration=123.458490378 podStartE2EDuration="2m3.458490378s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:22.456098861 +0000 UTC m=+145.472117603" watchObservedRunningTime="2025-10-10 09:13:22.458490378 +0000 UTC m=+145.474509120" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.511896 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.512270 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.012237664 +0000 UTC m=+146.028256406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.526371 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.527075 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.027058564 +0000 UTC m=+146.043077306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.542143 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfh8" podStartSLOduration=125.54211154 podStartE2EDuration="2m5.54211154s" podCreationTimestamp="2025-10-10 09:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:22.527461037 +0000 UTC m=+145.543479779" watchObservedRunningTime="2025-10-10 09:13:22.54211154 +0000 UTC m=+145.558130282" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.557062 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw"] Oct 10 09:13:22 crc kubenswrapper[4669]: W1010 09:13:22.595595 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod201e2060_b78b_4f0b_a9a7_6855bd9158eb.slice/crio-31a066a8de2d9ace83673bdbc15fad53a6c09e6d3c7bffb31e4b12413957b203 WatchSource:0}: Error finding container 31a066a8de2d9ace83673bdbc15fad53a6c09e6d3c7bffb31e4b12413957b203: Status 404 returned error can't find the container with id 31a066a8de2d9ace83673bdbc15fad53a6c09e6d3c7bffb31e4b12413957b203 Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.630376 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.630811 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.130793505 +0000 UTC m=+146.146812237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.631867 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.636106 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.732770 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.23274932 +0000 UTC m=+146.248768062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.731665 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.839199 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.839678 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.339659753 +0000 UTC m=+146.355678495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.943225 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:22 crc kubenswrapper[4669]: E1010 09:13:22.943602 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.4435692 +0000 UTC m=+146.459587942 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.997698 4669 patch_prober.go:28] interesting pod/apiserver-76f77b778f-x6svn container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]log ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]etcd ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/generic-apiserver-start-informers ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/max-in-flight-filter ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 10 09:13:22 crc kubenswrapper[4669]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 10 09:13:22 crc kubenswrapper[4669]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/project.openshift.io-projectcache ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 10 09:13:22 crc kubenswrapper[4669]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 10 09:13:22 crc kubenswrapper[4669]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 10 09:13:22 crc kubenswrapper[4669]: livez check failed Oct 10 09:13:22 crc kubenswrapper[4669]: I1010 09:13:22.997782 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" podUID="20e8f550-62f2-4d57-9f9f-62123ac8bc2e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.045154 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.045767 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.545744662 +0000 UTC m=+146.561763414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.046105 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.046629 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.546615649 +0000 UTC m=+146.562634391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.147819 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.148352 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.648331466 +0000 UTC m=+146.664350208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.186796 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" event={"ID":"033ee508-b7b5-47cf-984e-6d1c2e41b8ef","Type":"ContainerStarted","Data":"c16dfb086ef3aaa97a160251e0ec2e6dbf82da7b4454546fb58f633961313e7c"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.219936 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" event={"ID":"52821f37-1b46-4913-b0ac-c00e0a899193","Type":"ContainerStarted","Data":"542a13765a1b4246c68b8c5a74a7f5cd09ba046c1ebb7d1f1b463f898cecaf42"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.251746 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.252170 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.75215359 +0000 UTC m=+146.768172332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.273077 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" event={"ID":"24dc9ac3-3322-40f6-a99b-433280694265","Type":"ContainerStarted","Data":"a1ee77e12ad0fcec725b4e248b30ce0e8b8da8bcdbb286a89b7c8d66c5cd08e9"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.290489 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" event={"ID":"f273bc2e-26d7-40fe-9150-fb582fb14618","Type":"ContainerStarted","Data":"8d1f68aee4b2a0bf05c0666863f226fcc25564ab8eafe124963b338c49bdc624"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.330774 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" event={"ID":"521f5278-6996-4b67-a6ab-d46bd1a35acb","Type":"ContainerStarted","Data":"f868ec54d21ba1b4857820f6758a868a96f3edf8a4d07e83a792d83cf18282d1"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.331292 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" event={"ID":"521f5278-6996-4b67-a6ab-d46bd1a35acb","Type":"ContainerStarted","Data":"3796996a5b3adc42a7ae4de33c90c1eee0023f4e5783eeb14eeba2455d3cd3d6"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.353400 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.355095 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.855075005 +0000 UTC m=+146.871093757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.377909 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" event={"ID":"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550","Type":"ContainerStarted","Data":"aa5e118299cb95b8a03933c621a499749c7a3a116854226b99a004c7248fbd05"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.388635 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" event={"ID":"5783db5c-e7ff-4423-9188-f75344b4a111","Type":"ContainerStarted","Data":"55b5eb84ae66b19498ff632a81565b877d62d52f1e67b648eb77fdc4740ab461"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.404478 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.423662 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" event={"ID":"2f33e625-27c7-4380-be4c-8468bde0feb7","Type":"ContainerStarted","Data":"4fd51ea0776290f4c00e0b8b550ef010f878f8ffa1ecd0e9207fc080abe9183a"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.447773 4669 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-2jf4j container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.447838 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" podUID="15cafda9-1dc9-40b0-af56-024dfaf32754" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.456208 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.457951 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:23.957932298 +0000 UTC m=+146.973951040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.506331 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" event={"ID":"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4","Type":"ContainerStarted","Data":"fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.506391 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.514315 4669 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pb9vp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.514377 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" podUID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.558603 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.559905 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.059885463 +0000 UTC m=+147.075904205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.568932 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hsgp7" event={"ID":"d0b0ecc3-1f98-4345-8095-2a69aedeeac5","Type":"ContainerStarted","Data":"37b0828a69fcc735ef0e13c03d243a5dc86cdc62c80747f4682b2dc8c3fe0a85"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.663519 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.663895 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.163881193 +0000 UTC m=+147.179899935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.674298 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" event={"ID":"89852882-ff51-4b3c-9e60-8a862b7d18cc","Type":"ContainerStarted","Data":"d38c4887cee8b3cabd6236339340f1708d5bf9af678b9ab20a08ec260daf806b"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.675006 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.699616 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" event={"ID":"0e98983e-f7ef-4a56-a608-3715bc9f2ae7","Type":"ContainerStarted","Data":"b51fc267ec576ebed157809ab1521cd567936e1f40f8c9f617446e34005e0375"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.708441 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" event={"ID":"98efe68d-5663-488a-92ca-c02a288b524f","Type":"ContainerStarted","Data":"f8edf14c55d21fe2700b9025389717cb6da78032ec4ae2098d6d03ed8542851b"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.722879 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ngvr9" event={"ID":"3f266776-152e-4b8c-b4e4-98a578b5f238","Type":"ContainerStarted","Data":"cbd2d9a05ea922a531f64c7604800dca02ee623d8ee9ebf6ab3ecc316144661f"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.723202 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ngvr9" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.732656 4669 patch_prober.go:28] interesting pod/downloads-7954f5f757-ngvr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.732715 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ngvr9" podUID="3f266776-152e-4b8c-b4e4-98a578b5f238" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.737038 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" event={"ID":"201e2060-b78b-4f0b-a9a7-6855bd9158eb","Type":"ContainerStarted","Data":"31a066a8de2d9ace83673bdbc15fad53a6c09e6d3c7bffb31e4b12413957b203"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.760118 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rnpbk" event={"ID":"57749860-6c06-44f2-bbd9-382cf3bbd3c8","Type":"ContainerStarted","Data":"6e2eda4c10b24960300c285ebeb78174ee9fdf97f7c174cf2db24212f51865c7"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.767078 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.768618 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.268597936 +0000 UTC m=+147.284616678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.792122 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-llrlv" event={"ID":"6ec9ad95-b8ea-459b-a151-1601edf706ea","Type":"ContainerStarted","Data":"7018e5c6eb57e0e0940267fc38f1620b8c7e5c87ba10212d0c3cfe47056d9bf2"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.874390 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4xs8s" podStartSLOduration=124.874366033 podStartE2EDuration="2m4.874366033s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:23.812106842 +0000 UTC m=+146.828125584" watchObservedRunningTime="2025-10-10 09:13:23.874366033 +0000 UTC m=+146.890384775" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.875847 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w5qk7" podStartSLOduration=124.875841811 podStartE2EDuration="2m4.875841811s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:23.87394734 +0000 UTC m=+146.889966082" watchObservedRunningTime="2025-10-10 09:13:23.875841811 +0000 UTC m=+146.891860553" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.879488 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.879871 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.379854101 +0000 UTC m=+147.395872843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.919247 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" podStartSLOduration=125.919222363 podStartE2EDuration="2m5.919222363s" podCreationTimestamp="2025-10-10 09:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:23.903274827 +0000 UTC m=+146.919293569" watchObservedRunningTime="2025-10-10 09:13:23.919222363 +0000 UTC m=+146.935241105" Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.942014 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" event={"ID":"d3c7621d-224f-4ada-954c-5f9f051a6132","Type":"ContainerStarted","Data":"57ae898a7505774d35438f7b61406bfe1e78deb4dffec56cf23ee6d9afd93b78"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.977309 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" event={"ID":"e57d53b2-5487-4820-aa61-9070eb866eae","Type":"ContainerStarted","Data":"e49ca81ba9f1598af5d13479754177409e336b62e2e11330cac43a2116fec639"} Oct 10 09:13:23 crc kubenswrapper[4669]: I1010 09:13:23.983142 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:23 crc kubenswrapper[4669]: E1010 09:13:23.983636 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.483612983 +0000 UTC m=+147.499631725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.021333 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" event={"ID":"2a92e481-9a5c-4238-988f-f2d0b852ebca","Type":"ContainerStarted","Data":"8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f"} Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.022530 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.075412 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" event={"ID":"8f9070f5-5a97-4cfd-9eea-9c5c011ea064","Type":"ContainerStarted","Data":"9214e512ebb5bbf48f7ff5c4d69b1e52924456fe910a69bab70905effc06089c"} Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.088706 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.097924 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" event={"ID":"c2151912-a994-4992-915d-d737497335d7","Type":"ContainerStarted","Data":"df651f87359185e37afc6d7057f3be73b14e6660df11280c2774413b08c679af"} Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.098959 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.102899 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.602881666 +0000 UTC m=+147.618900398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.104712 4669 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kdgw8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" start-of-body= Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.104757 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" podUID="c2151912-a994-4992-915d-d737497335d7" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.119967 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" event={"ID":"7b4d82f5-1059-46f6-a639-664fd1e47963","Type":"ContainerStarted","Data":"4e5d03e0e8b30dc15ab32ca4f26f125f4de263db611b337a56aa27bc3b0574ab"} Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.142701 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" event={"ID":"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a","Type":"ContainerStarted","Data":"ec13e9eb418c1346670ff7a0da8b330e5a72ee6d4fc90e0cb509f9799364df7b"} Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.151365 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-lvq4s" podStartSLOduration=125.151352452 podStartE2EDuration="2m5.151352452s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:23.978949212 +0000 UTC m=+146.994967954" watchObservedRunningTime="2025-10-10 09:13:24.151352452 +0000 UTC m=+147.167371194" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.165852 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" event={"ID":"6db04d09-40ad-4a7b-81fd-9d2ebbb1d1e1","Type":"ContainerStarted","Data":"fcaf1e619640425fe94d0a337d11cb0bf7f196ac0468cf5d49e273f1e0de5e70"} Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.170216 4669 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nvfcj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.170276 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" podUID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.190059 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.190251 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.690215188 +0000 UTC m=+147.706233990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.191141 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.197008 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.696981747 +0000 UTC m=+147.713000639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.251533 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-nbx9n" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.290521 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.290602 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.292361 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.293838 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.793817225 +0000 UTC m=+147.809835967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.309732 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" podStartSLOduration=125.309703848 podStartE2EDuration="2m5.309703848s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.152609362 +0000 UTC m=+147.168628104" watchObservedRunningTime="2025-10-10 09:13:24.309703848 +0000 UTC m=+147.325722590" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.315958 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" podStartSLOduration=125.31593635 podStartE2EDuration="2m5.31593635s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.30448399 +0000 UTC m=+147.320502722" watchObservedRunningTime="2025-10-10 09:13:24.31593635 +0000 UTC m=+147.331955092" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.398495 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.398921 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:24.89890436 +0000 UTC m=+147.914923102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.436616 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-f6dtf" podStartSLOduration=125.436598938 podStartE2EDuration="2m5.436598938s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.425442908 +0000 UTC m=+147.441461650" watchObservedRunningTime="2025-10-10 09:13:24.436598938 +0000 UTC m=+147.452617680" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.439610 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" podStartSLOduration=125.439599745 podStartE2EDuration="2m5.439599745s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.372090244 +0000 UTC m=+147.388108986" watchObservedRunningTime="2025-10-10 09:13:24.439599745 +0000 UTC m=+147.455618487" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.505682 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.506910 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.006885219 +0000 UTC m=+148.022903961 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.544139 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ngvr9" podStartSLOduration=125.544122413 podStartE2EDuration="2m5.544122413s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.480574619 +0000 UTC m=+147.496593361" watchObservedRunningTime="2025-10-10 09:13:24.544122413 +0000 UTC m=+147.560141155" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.607321 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.607769 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.107752778 +0000 UTC m=+148.123771530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.623095 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-4mpz6" podStartSLOduration=125.623073013 podStartE2EDuration="2m5.623073013s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.542851271 +0000 UTC m=+147.558870013" watchObservedRunningTime="2025-10-10 09:13:24.623073013 +0000 UTC m=+147.639091755" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.623253 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" podStartSLOduration=125.623249018 podStartE2EDuration="2m5.623249018s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.621017356 +0000 UTC m=+147.637036118" watchObservedRunningTime="2025-10-10 09:13:24.623249018 +0000 UTC m=+147.639267760" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.711184 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.711756 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.211737958 +0000 UTC m=+148.227756700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.712686 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9lpfb" podStartSLOduration=125.712664187 podStartE2EDuration="2m5.712664187s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.70688105 +0000 UTC m=+147.722899792" watchObservedRunningTime="2025-10-10 09:13:24.712664187 +0000 UTC m=+147.728682929" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.768888 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" podStartSLOduration=125.768854883 podStartE2EDuration="2m5.768854883s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.75946254 +0000 UTC m=+147.775481282" watchObservedRunningTime="2025-10-10 09:13:24.768854883 +0000 UTC m=+147.784873625" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.813527 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.814304 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.31428127 +0000 UTC m=+148.330300012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.876817 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-xk9nh" podStartSLOduration=125.87678691 podStartE2EDuration="2m5.87678691s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:24.875147847 +0000 UTC m=+147.891166589" watchObservedRunningTime="2025-10-10 09:13:24.87678691 +0000 UTC m=+147.892805642" Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.914750 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.914858 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.414839099 +0000 UTC m=+148.430857841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:24 crc kubenswrapper[4669]: I1010 09:13:24.915097 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:24 crc kubenswrapper[4669]: E1010 09:13:24.915371 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.415364286 +0000 UTC m=+148.431383028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.016786 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.017274 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.517255619 +0000 UTC m=+148.533274361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.022623 4669 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-l67bg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.022670 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" podUID="2a92e481-9a5c-4238-988f-f2d0b852ebca" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.118411 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.118845 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.61882522 +0000 UTC m=+148.634843962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.169009 4669 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-kkf8r container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": context deadline exceeded" start-of-body= Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.169144 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" podUID="55a8f8c1-7a6c-474f-b422-60d886882a59" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": context deadline exceeded" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.171827 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" event={"ID":"8f9070f5-5a97-4cfd-9eea-9c5c011ea064","Type":"ContainerStarted","Data":"c36f17e22fa512804f49759d1c534dc0285f1bd36ea4fbf77e8fa379c2cd5f07"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.172976 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" event={"ID":"7b4d82f5-1059-46f6-a639-664fd1e47963","Type":"ContainerStarted","Data":"b9faf6294c6feaead1aa12f1f400f9698035465e2988a6d29d5f472ab2d680ea"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.174492 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" event={"ID":"52821f37-1b46-4913-b0ac-c00e0a899193","Type":"ContainerStarted","Data":"fc9e3ddf6ea1467e37e3daa625cbd5bd094ab208bd8be23af17883de40271132"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.175632 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" event={"ID":"c62d6a28-def0-41d5-a931-3bb27183deef","Type":"ContainerStarted","Data":"79e270671144bc96ed859de60a551d7c094a8bc36dad11baeb7a59da40773e8e"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.176204 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.177167 4669 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-rwgqq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.177208 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" podUID="c62d6a28-def0-41d5-a931-3bb27183deef" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.178514 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" event={"ID":"182c02d0-fdb0-4a01-bb6f-efdc3bbcb550","Type":"ContainerStarted","Data":"f8ed976dc04d050b45d0810cbb130b4abceb4c15cf7eb239fc049a62463776d5"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.179818 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" event={"ID":"15cafda9-1dc9-40b0-af56-024dfaf32754","Type":"ContainerStarted","Data":"60ecd659d1ed2fc6419654441fb4e6324c7b4667bcd3fa2b9b2f5ab185ca685e"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.182034 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hsgp7" event={"ID":"d0b0ecc3-1f98-4345-8095-2a69aedeeac5","Type":"ContainerStarted","Data":"a14abc60bebf33ea950b815e43858d24ee2935eabb5063d9421abbd227dbfb0b"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.183512 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" event={"ID":"2f33e625-27c7-4380-be4c-8468bde0feb7","Type":"ContainerStarted","Data":"d520a0bcd6a046c31bac9a29e94a5e4876454547b793de46e53637145e768391"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.183537 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" event={"ID":"2f33e625-27c7-4380-be4c-8468bde0feb7","Type":"ContainerStarted","Data":"b347b3854ec8efda5f08e504937831a47299971dbc1f9e2659e4a5612b6c61f0"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.184767 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-llrlv" event={"ID":"6ec9ad95-b8ea-459b-a151-1601edf706ea","Type":"ContainerStarted","Data":"5e5360ef8f7de82a7614ee9d6194bb3273b56b18372d9bb222c869eb9920f29b"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.186311 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" event={"ID":"acbd17d0-d2e2-4a68-991c-3553eb69f8ea","Type":"ContainerStarted","Data":"aa1a62f0e69a817bd8ef5404ea0a0489bc75cc6451b2fca16850117477451963"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.186353 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" event={"ID":"acbd17d0-d2e2-4a68-991c-3553eb69f8ea","Type":"ContainerStarted","Data":"227b1045c2e08378a0c4459d18c4a115e20040e2654e2f25fd098089f0fa46a4"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.187706 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zqhsj" event={"ID":"e505e6fd-e00f-40a8-a575-6102ec52059d","Type":"ContainerStarted","Data":"670c862014a9b6fe2bd7ca87d69f62e206d393f62678d2917d7a3c438f1c1e9e"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.189479 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" event={"ID":"e5962f59-9740-46e8-9866-a57806721845","Type":"ContainerStarted","Data":"80dcd78feacddcf570fdfa3acad5458ce1f2073155633d73c068a73c8b1c18ca"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.190338 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" event={"ID":"caf4f8a4-2c97-48b9-9fd9-ff717ef25b8a","Type":"ContainerStarted","Data":"78229e079fa122546333309ade1cfa3789e27f550bcad59b78d751e0f6e14287"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.192109 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" event={"ID":"5783db5c-e7ff-4423-9188-f75344b4a111","Type":"ContainerStarted","Data":"58ae92d08c9443112d2bc0f37d3465b730b9499103dc17e630c11910a784808d"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.194769 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rnpbk" event={"ID":"57749860-6c06-44f2-bbd9-382cf3bbd3c8","Type":"ContainerStarted","Data":"c20ea9a16d55095ba1fd6f765122cbd181b1a010724156cdad54a76d0e72e30a"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.194804 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rnpbk" event={"ID":"57749860-6c06-44f2-bbd9-382cf3bbd3c8","Type":"ContainerStarted","Data":"f6e10039955295165d1e0c52679568e8a16c375b82b2bdc01170fac8ab698775"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.195176 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.197565 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" event={"ID":"d3c7621d-224f-4ada-954c-5f9f051a6132","Type":"ContainerStarted","Data":"0ec44b664471ac4626e3c69e861439c2a00024da1cacebbb50e036f8c77a66b6"} Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.198473 4669 patch_prober.go:28] interesting pod/downloads-7954f5f757-ngvr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.198519 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ngvr9" podUID="3f266776-152e-4b8c-b4e4-98a578b5f238" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.217536 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.219224 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.219862 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.220185 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.720159344 +0000 UTC m=+148.736178086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.221156 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.224112 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.724092851 +0000 UTC m=+148.740111593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.232303 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2jf4j" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.233712 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.251738 4669 patch_prober.go:28] interesting pod/router-default-5444994796-zqhsj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 09:13:25 crc kubenswrapper[4669]: [-]has-synced failed: reason withheld Oct 10 09:13:25 crc kubenswrapper[4669]: [+]process-running ok Oct 10 09:13:25 crc kubenswrapper[4669]: healthz check failed Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.251806 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zqhsj" podUID="e505e6fd-e00f-40a8-a575-6102ec52059d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.321910 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.324919 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.824885658 +0000 UTC m=+148.840904400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.360219 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.378432 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-stkm9" podStartSLOduration=126.378395407 podStartE2EDuration="2m6.378395407s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.277736284 +0000 UTC m=+148.293755046" watchObservedRunningTime="2025-10-10 09:13:25.378395407 +0000 UTC m=+148.394414149" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.409460 4669 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-kkf8r container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.409569 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kkf8r" podUID="55a8f8c1-7a6c-474f-b422-60d886882a59" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.425951 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.426499 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:25.926479539 +0000 UTC m=+148.942498281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.433741 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rcqcc" podStartSLOduration=126.433715814 podStartE2EDuration="2m6.433715814s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.369468988 +0000 UTC m=+148.385487730" watchObservedRunningTime="2025-10-10 09:13:25.433715814 +0000 UTC m=+148.449734556" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.447895 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-klgpb" podStartSLOduration=126.447863071 podStartE2EDuration="2m6.447863071s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.434697955 +0000 UTC m=+148.450716697" watchObservedRunningTime="2025-10-10 09:13:25.447863071 +0000 UTC m=+148.463881813" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.495554 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-8226f" podStartSLOduration=126.495532981 podStartE2EDuration="2m6.495532981s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.490922951 +0000 UTC m=+148.506941693" watchObservedRunningTime="2025-10-10 09:13:25.495532981 +0000 UTC m=+148.511551723" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.528001 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.528547 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.028526587 +0000 UTC m=+149.044545329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.630248 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.630304 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.630336 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.631387 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.131361479 +0000 UTC m=+149.147380391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.632421 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.638231 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.656351 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" podStartSLOduration=126.656331146 podStartE2EDuration="2m6.656331146s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.612100127 +0000 UTC m=+148.628118869" watchObservedRunningTime="2025-10-10 09:13:25.656331146 +0000 UTC m=+148.672349888" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.716339 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r6blz" podStartSLOduration=126.716323004 podStartE2EDuration="2m6.716323004s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.713777162 +0000 UTC m=+148.729795894" watchObservedRunningTime="2025-10-10 09:13:25.716323004 +0000 UTC m=+148.732341746" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.716815 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j89vb" podStartSLOduration=126.71681032 podStartE2EDuration="2m6.71681032s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.65862904 +0000 UTC m=+148.674647782" watchObservedRunningTime="2025-10-10 09:13:25.71681032 +0000 UTC m=+148.732829062" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.731763 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.731952 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.732037 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.733526 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.233479969 +0000 UTC m=+149.249498711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.749405 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.750127 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.752610 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.831029 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hsgp7" podStartSLOduration=9.83101182 podStartE2EDuration="9.83101182s" podCreationTimestamp="2025-10-10 09:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.829965086 +0000 UTC m=+148.845983818" watchObservedRunningTime="2025-10-10 09:13:25.83101182 +0000 UTC m=+148.847030562" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.833482 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.834075 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.334062508 +0000 UTC m=+149.350081240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.887407 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-d7jk5" podStartSLOduration=126.887377971 podStartE2EDuration="2m6.887377971s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.88551055 +0000 UTC m=+148.901529292" watchObservedRunningTime="2025-10-10 09:13:25.887377971 +0000 UTC m=+148.903396703" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.940250 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:25 crc kubenswrapper[4669]: E1010 09:13:25.940661 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.440641392 +0000 UTC m=+149.456660134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.994507 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-llrlv" podStartSLOduration=9.994486091 podStartE2EDuration="9.994486091s" podCreationTimestamp="2025-10-10 09:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.993039235 +0000 UTC m=+149.009057977" watchObservedRunningTime="2025-10-10 09:13:25.994486091 +0000 UTC m=+149.010504833" Oct 10 09:13:25 crc kubenswrapper[4669]: I1010 09:13:25.995423 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-zqhsj" podStartSLOduration=126.995416981 podStartE2EDuration="2m6.995416981s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:25.95792694 +0000 UTC m=+148.973945682" watchObservedRunningTime="2025-10-10 09:13:25.995416981 +0000 UTC m=+149.011435723" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.022333 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.035888 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.041582 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.041902 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.541890133 +0000 UTC m=+149.557908875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.145138 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.145513 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.64546977 +0000 UTC m=+149.661488512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.145671 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.146145 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.64612777 +0000 UTC m=+149.662146512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.196976 4669 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pb9vp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.197046 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" podUID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.198773 4669 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kdgw8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.198811 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" podUID="c2151912-a994-4992-915d-d737497335d7" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.210390 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" event={"ID":"201e2060-b78b-4f0b-a9a7-6855bd9158eb","Type":"ContainerStarted","Data":"a521ae1342ffc692c73dafc5b2b0fa174a1b4b0edf4cc22a22fec6dc011342fd"} Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.229430 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rwgqq" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.233753 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" podStartSLOduration=127.233732221 podStartE2EDuration="2m7.233732221s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:26.189461661 +0000 UTC m=+149.205480403" watchObservedRunningTime="2025-10-10 09:13:26.233732221 +0000 UTC m=+149.249750963" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.238783 4669 patch_prober.go:28] interesting pod/router-default-5444994796-zqhsj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 09:13:26 crc kubenswrapper[4669]: [-]has-synced failed: reason withheld Oct 10 09:13:26 crc kubenswrapper[4669]: [+]process-running ok Oct 10 09:13:26 crc kubenswrapper[4669]: healthz check failed Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.238838 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zqhsj" podUID="e505e6fd-e00f-40a8-a575-6102ec52059d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.250769 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.252569 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.752547458 +0000 UTC m=+149.768566200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.254853 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.267083 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.767063898 +0000 UTC m=+149.783082640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.356839 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.357526 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.85750087 +0000 UTC m=+149.873519612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.370333 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-xvhtj" podStartSLOduration=127.370314434 podStartE2EDuration="2m7.370314434s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:26.368321239 +0000 UTC m=+149.384339981" watchObservedRunningTime="2025-10-10 09:13:26.370314434 +0000 UTC m=+149.386333176" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.375751 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rnpbk" podStartSLOduration=10.375727288 podStartE2EDuration="10.375727288s" podCreationTimestamp="2025-10-10 09:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:26.275116898 +0000 UTC m=+149.291135630" watchObservedRunningTime="2025-10-10 09:13:26.375727288 +0000 UTC m=+149.391746030" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.461328 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.461724 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:26.961707887 +0000 UTC m=+149.977726629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.566350 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.566869 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.066851334 +0000 UTC m=+150.082870076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.667313 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.667662 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.16764947 +0000 UTC m=+150.183668212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.692217 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-grdvw" podStartSLOduration=127.692198313 podStartE2EDuration="2m7.692198313s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:26.580830395 +0000 UTC m=+149.596849137" watchObservedRunningTime="2025-10-10 09:13:26.692198313 +0000 UTC m=+149.708217055" Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.768077 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.768305 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.268287742 +0000 UTC m=+150.284306484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.869065 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.869554 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.369534473 +0000 UTC m=+150.385553215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:26 crc kubenswrapper[4669]: I1010 09:13:26.970916 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:26 crc kubenswrapper[4669]: E1010 09:13:26.971155 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.471137736 +0000 UTC m=+150.487156478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.073334 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.073692 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.573679158 +0000 UTC m=+150.589697900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.183044 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.183232 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.683200447 +0000 UTC m=+150.699219189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.183596 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.183867 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.683859659 +0000 UTC m=+150.699878401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.211770 4669 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kdgw8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.211835 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" podUID="c2151912-a994-4992-915d-d737497335d7" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.237797 4669 patch_prober.go:28] interesting pod/router-default-5444994796-zqhsj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 09:13:27 crc kubenswrapper[4669]: [-]has-synced failed: reason withheld Oct 10 09:13:27 crc kubenswrapper[4669]: [+]process-running ok Oct 10 09:13:27 crc kubenswrapper[4669]: healthz check failed Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.237852 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zqhsj" podUID="e505e6fd-e00f-40a8-a575-6102ec52059d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.238801 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" event={"ID":"201e2060-b78b-4f0b-a9a7-6855bd9158eb","Type":"ContainerStarted","Data":"9215b5f72cdfb13196197b124bf4f763564ef65de76d8154bd2bd0a0690596da"} Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.284665 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.285664 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.785645137 +0000 UTC m=+150.801663879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.384771 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d7z4x"] Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.386989 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.387340 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.887323802 +0000 UTC m=+150.903342544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.387653 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.398062 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.430133 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7z4x"] Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.489117 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.489223 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.989205394 +0000 UTC m=+151.005224136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.489372 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.489418 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-catalog-content\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.489450 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-762w7\" (UniqueName: \"kubernetes.io/projected/f38b0724-9f7b-4108-926f-e2192c26c49e-kube-api-access-762w7\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.489474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-utilities\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.489749 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:27.98974195 +0000 UTC m=+151.005760692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: W1010 09:13:27.508465 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-89d3bcba42df79ca878d161f8c2c13ed70b5cd1ec7eb336fca42d9261a9d016b WatchSource:0}: Error finding container 89d3bcba42df79ca878d161f8c2c13ed70b5cd1ec7eb336fca42d9261a9d016b: Status 404 returned error can't find the container with id 89d3bcba42df79ca878d161f8c2c13ed70b5cd1ec7eb336fca42d9261a9d016b Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.553931 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-csltr"] Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.555735 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.558420 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.590196 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.590512 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-catalog-content\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.590560 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-762w7\" (UniqueName: \"kubernetes.io/projected/f38b0724-9f7b-4108-926f-e2192c26c49e-kube-api-access-762w7\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.590601 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-utilities\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.590977 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.09094525 +0000 UTC m=+151.106963992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.591073 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-utilities\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.591411 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-catalog-content\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.641912 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.646522 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-x6svn" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.662888 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-762w7\" (UniqueName: \"kubernetes.io/projected/f38b0724-9f7b-4108-926f-e2192c26c49e-kube-api-access-762w7\") pod \"community-operators-d7z4x\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.678014 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.678053 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.682068 4669 patch_prober.go:28] interesting pod/console-f9d7485db-wqc7t container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.682112 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wqc7t" podUID="69730259-7614-4869-bc84-10e17d8d13df" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.693005 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-catalog-content\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.693047 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsfmq\" (UniqueName: \"kubernetes.io/projected/85775a5b-bc26-43ba-a512-544217dec697-kube-api-access-vsfmq\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.693073 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.693105 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-utilities\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.693446 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.193434542 +0000 UTC m=+151.209453274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.712088 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.774927 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csltr"] Oct 10 09:13:27 crc kubenswrapper[4669]: W1010 09:13:27.780390 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-90a4896ba10ac5cda9b492d0050dda69151f4ca1f722d593416b9ff0fa08a7c7 WatchSource:0}: Error finding container 90a4896ba10ac5cda9b492d0050dda69151f4ca1f722d593416b9ff0fa08a7c7: Status 404 returned error can't find the container with id 90a4896ba10ac5cda9b492d0050dda69151f4ca1f722d593416b9ff0fa08a7c7 Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.794153 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.794344 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-utilities\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.794568 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-catalog-content\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.794609 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsfmq\" (UniqueName: \"kubernetes.io/projected/85775a5b-bc26-43ba-a512-544217dec697-kube-api-access-vsfmq\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.794866 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.294850218 +0000 UTC m=+151.310868960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.803177 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-utilities\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.814561 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-catalog-content\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.864147 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cswwd"] Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.872648 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsfmq\" (UniqueName: \"kubernetes.io/projected/85775a5b-bc26-43ba-a512-544217dec697-kube-api-access-vsfmq\") pod \"certified-operators-csltr\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.876242 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.896890 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.897842 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:27 crc kubenswrapper[4669]: E1010 09:13:27.898158 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.398146965 +0000 UTC m=+151.414165707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.926976 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cswwd"] Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.976680 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wlqpr"] Oct 10 09:13:27 crc kubenswrapper[4669]: I1010 09:13:27.977628 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.004500 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.004716 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.504695538 +0000 UTC m=+151.520714280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.005216 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.005280 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-utilities\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.005310 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b56k\" (UniqueName: \"kubernetes.io/projected/efc3a420-b563-4f1b-8d25-17d363007b8d-kube-api-access-8b56k\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.005339 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-catalog-content\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.006058 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.506049462 +0000 UTC m=+151.522068204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.106297 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.106698 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7zwz\" (UniqueName: \"kubernetes.io/projected/42e85702-8412-4598-beb9-0e0ab5ae944e-kube-api-access-w7zwz\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.106757 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-utilities\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.106786 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b56k\" (UniqueName: \"kubernetes.io/projected/efc3a420-b563-4f1b-8d25-17d363007b8d-kube-api-access-8b56k\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.106818 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-catalog-content\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.106848 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-catalog-content\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.106880 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-utilities\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.107029 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.607006084 +0000 UTC m=+151.623024826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.107392 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-utilities\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.107884 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-catalog-content\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.113379 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlqpr"] Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.189248 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b56k\" (UniqueName: \"kubernetes.io/projected/efc3a420-b563-4f1b-8d25-17d363007b8d-kube-api-access-8b56k\") pod \"community-operators-cswwd\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.208675 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-catalog-content\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.209110 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-utilities\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.209174 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.209200 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7zwz\" (UniqueName: \"kubernetes.io/projected/42e85702-8412-4598-beb9-0e0ab5ae944e-kube-api-access-w7zwz\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.209970 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-catalog-content\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.213193 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.712979438 +0000 UTC m=+151.728998180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.213647 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.240926 4669 patch_prober.go:28] interesting pod/router-default-5444994796-zqhsj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 09:13:28 crc kubenswrapper[4669]: [-]has-synced failed: reason withheld Oct 10 09:13:28 crc kubenswrapper[4669]: [+]process-running ok Oct 10 09:13:28 crc kubenswrapper[4669]: healthz check failed Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.240981 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zqhsj" podUID="e505e6fd-e00f-40a8-a575-6102ec52059d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.249370 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7zwz\" (UniqueName: \"kubernetes.io/projected/42e85702-8412-4598-beb9-0e0ab5ae944e-kube-api-access-w7zwz\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.249490 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-utilities\") pod \"certified-operators-wlqpr\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.280127 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"90a4896ba10ac5cda9b492d0050dda69151f4ca1f722d593416b9ff0fa08a7c7"} Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.297635 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"89d3bcba42df79ca878d161f8c2c13ed70b5cd1ec7eb336fca42d9261a9d016b"} Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.305884 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.310897 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"fa04184708734751bc50a7a55543fed6f30a4c49ac30aafe086ab0657ecb5fcc"} Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.311319 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.311855 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.811835122 +0000 UTC m=+151.827853864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.413629 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.415761 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:28.915748709 +0000 UTC m=+151.931767451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.529161 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.529659 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.029641439 +0000 UTC m=+152.045660181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.633102 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.633705 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.13369107 +0000 UTC m=+152.149709812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.734058 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.734407 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.234389594 +0000 UTC m=+152.250408336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.821680 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.822776 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.836240 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.836623 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.336610426 +0000 UTC m=+152.352629158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.847313 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.921175 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:13:28 crc kubenswrapper[4669]: I1010 09:13:28.938622 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:28 crc kubenswrapper[4669]: E1010 09:13:28.942497 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.442471576 +0000 UTC m=+152.458490318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.040341 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.040920 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cswwd"] Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.042148 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.542134797 +0000 UTC m=+152.558153539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.062223 4669 patch_prober.go:28] interesting pod/downloads-7954f5f757-ngvr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.062282 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ngvr9" podUID="3f266776-152e-4b8c-b4e4-98a578b5f238" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.062388 4669 patch_prober.go:28] interesting pod/downloads-7954f5f757-ngvr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.062439 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ngvr9" podUID="3f266776-152e-4b8c-b4e4-98a578b5f238" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.142546 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.142939 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.642922563 +0000 UTC m=+152.658941305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.153887 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.163999 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-csltr"] Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.244286 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.245551 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.745539458 +0000 UTC m=+152.761558200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.246744 4669 patch_prober.go:28] interesting pod/router-default-5444994796-zqhsj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 09:13:29 crc kubenswrapper[4669]: [-]has-synced failed: reason withheld Oct 10 09:13:29 crc kubenswrapper[4669]: [+]process-running ok Oct 10 09:13:29 crc kubenswrapper[4669]: healthz check failed Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.246795 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zqhsj" podUID="e505e6fd-e00f-40a8-a575-6102ec52059d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.328505 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2c68a4e04d8cd883bc9cd18ee08be29fc33430ac1ab85172c3ce5a817f1bf5af"} Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.328793 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.330615 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cswwd" event={"ID":"efc3a420-b563-4f1b-8d25-17d363007b8d","Type":"ContainerStarted","Data":"dea38097e0a6b5751e80bc820cc7e63b674cc84b62b44dc5c3273bf0053e606b"} Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.334391 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csltr" event={"ID":"85775a5b-bc26-43ba-a512-544217dec697","Type":"ContainerStarted","Data":"8cc6f8647c0379240841e04538b4e7a1eee1a87c278fdcc4854afefea6573ea2"} Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.344167 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"08620f8db583a3c278079dcca71ea3b10703584f3975d8bb996677dc8254d39b"} Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.345060 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.345312 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.845297022 +0000 UTC m=+152.861315764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.360076 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7z4x"] Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.360515 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a34637ebf05fc52fc5776567be42deb645f7b1bed60c7a864d04b5f4bfbbf835"} Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.376127 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" event={"ID":"201e2060-b78b-4f0b-a9a7-6855bd9158eb","Type":"ContainerStarted","Data":"672f4e94dbce81b530ba7bffadb9c21d6ca09886192e18e2e697dffa873d02ab"} Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.390207 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zltk9" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.446875 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.447387 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:29.947373219 +0000 UTC m=+152.963391961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.536926 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlqpr"] Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.548409 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.548621 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.048577259 +0000 UTC m=+153.064596001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.548754 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.549434 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.049423427 +0000 UTC m=+153.065442239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.555333 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mwjd4"] Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.556339 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.558767 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.612864 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kdgw8" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.641659 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwjd4"] Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.650017 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.650255 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-catalog-content\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.650556 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnnwv\" (UniqueName: \"kubernetes.io/projected/bf327e89-3302-463b-aaaf-87391e96aae1-kube-api-access-dnnwv\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.650681 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-utilities\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.650794 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.15077635 +0000 UTC m=+153.166795092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.752403 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnnwv\" (UniqueName: \"kubernetes.io/projected/bf327e89-3302-463b-aaaf-87391e96aae1-kube-api-access-dnnwv\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.752489 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-utilities\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.752544 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-catalog-content\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.752602 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.752991 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.252975813 +0000 UTC m=+153.268994565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.756813 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-catalog-content\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.757419 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-utilities\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.853445 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.853825 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.35379832 +0000 UTC m=+153.369817072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.879608 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnnwv\" (UniqueName: \"kubernetes.io/projected/bf327e89-3302-463b-aaaf-87391e96aae1-kube-api-access-dnnwv\") pod \"redhat-marketplace-mwjd4\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.889714 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.946259 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-26c6v"] Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.947531 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:29 crc kubenswrapper[4669]: I1010 09:13:29.956399 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:29 crc kubenswrapper[4669]: E1010 09:13:29.956711 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.456698965 +0000 UTC m=+153.472717707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.047049 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-26c6v"] Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.057509 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.057734 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.557705348 +0000 UTC m=+153.573724090 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.057788 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-catalog-content\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.057974 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h2fr\" (UniqueName: \"kubernetes.io/projected/434cae8a-10fb-41c0-af9d-4756b10fc8cf-kube-api-access-7h2fr\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.058199 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-utilities\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.058282 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.058546 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.558533235 +0000 UTC m=+153.574551977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.161055 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.161343 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.661299925 +0000 UTC m=+153.677318667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.161505 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-utilities\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.161548 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.161601 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-catalog-content\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.161637 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h2fr\" (UniqueName: \"kubernetes.io/projected/434cae8a-10fb-41c0-af9d-4756b10fc8cf-kube-api-access-7h2fr\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.162125 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-utilities\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.162172 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.662160473 +0000 UTC m=+153.678179215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.162388 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-catalog-content\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.180450 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h2fr\" (UniqueName: \"kubernetes.io/projected/434cae8a-10fb-41c0-af9d-4756b10fc8cf-kube-api-access-7h2fr\") pod \"redhat-marketplace-26c6v\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.232940 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.244060 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.263490 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.263934 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.76391474 +0000 UTC m=+153.779933482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.301518 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.364831 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.365754 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.86574198 +0000 UTC m=+153.881760722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.390500 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" event={"ID":"201e2060-b78b-4f0b-a9a7-6855bd9158eb","Type":"ContainerStarted","Data":"de5ba5233503473fb53d8bc7c81ac082e1f3692e1d9e4e734f0f71280c67f2b8"} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.417446 4669 generic.go:334] "Generic (PLEG): container finished" podID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerID="b99a26a93642b2e3be49a6c4d75a910221e80209fa06c9048b352835d75ed127" exitCode=0 Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.417554 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cswwd" event={"ID":"efc3a420-b563-4f1b-8d25-17d363007b8d","Type":"ContainerDied","Data":"b99a26a93642b2e3be49a6c4d75a910221e80209fa06c9048b352835d75ed127"} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.417614 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-mfjcd" podStartSLOduration=14.417596506 podStartE2EDuration="14.417596506s" podCreationTimestamp="2025-10-10 09:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:30.416902903 +0000 UTC m=+153.432921655" watchObservedRunningTime="2025-10-10 09:13:30.417596506 +0000 UTC m=+153.433615248" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.420025 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.431817 4669 generic.go:334] "Generic (PLEG): container finished" podID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerID="4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331" exitCode=0 Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.431938 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlqpr" event={"ID":"42e85702-8412-4598-beb9-0e0ab5ae944e","Type":"ContainerDied","Data":"4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331"} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.431972 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlqpr" event={"ID":"42e85702-8412-4598-beb9-0e0ab5ae944e","Type":"ContainerStarted","Data":"7f74adaaec39cd3d56df61b81380455b8128ec4d9a4a0b41f7829781f99791df"} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.449770 4669 generic.go:334] "Generic (PLEG): container finished" podID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerID="a2ff080160fbb8f27fc7f3a8dacd594a5ebed97ee77c3f4822bf7bf07785c424" exitCode=0 Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.449847 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7z4x" event={"ID":"f38b0724-9f7b-4108-926f-e2192c26c49e","Type":"ContainerDied","Data":"a2ff080160fbb8f27fc7f3a8dacd594a5ebed97ee77c3f4822bf7bf07785c424"} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.449874 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7z4x" event={"ID":"f38b0724-9f7b-4108-926f-e2192c26c49e","Type":"ContainerStarted","Data":"9446e5e8490d001a695c2c6ae317ccaa097bd81da2c47bb3adf8d6178055b398"} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.465002 4669 generic.go:334] "Generic (PLEG): container finished" podID="85775a5b-bc26-43ba-a512-544217dec697" containerID="eba304693b0777d88678a70a67b6ee4dc7e0a865d39a43485e7b4c96f8980990" exitCode=0 Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.466053 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csltr" event={"ID":"85775a5b-bc26-43ba-a512-544217dec697","Type":"ContainerDied","Data":"eba304693b0777d88678a70a67b6ee4dc7e0a865d39a43485e7b4c96f8980990"} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.467173 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.468060 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:30.968043035 +0000 UTC m=+153.984061777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.470759 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-zqhsj" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.484054 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwjd4"] Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.518396 4669 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.574840 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.578843 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:31.078830245 +0000 UTC m=+154.094848987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.680854 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.681265 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:31.181243283 +0000 UTC m=+154.197262025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.724304 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p9txp"] Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.725543 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.729171 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.730920 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p9txp"] Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.782392 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.783033 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:31.283014472 +0000 UTC m=+154.299033214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.802500 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-26c6v"] Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.883787 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.883963 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:31.383923092 +0000 UTC m=+154.399941834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.884051 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-catalog-content\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.884083 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-utilities\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.884111 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmkkk\" (UniqueName: \"kubernetes.io/projected/cf42d4f9-7d35-43d2-a115-c82b5386be00-kube-api-access-lmkkk\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.884177 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.884519 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 09:13:31.384511061 +0000 UTC m=+154.400529803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w8ngm" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.985218 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.985471 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-catalog-content\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.985506 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-utilities\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.985538 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmkkk\" (UniqueName: \"kubernetes.io/projected/cf42d4f9-7d35-43d2-a115-c82b5386be00-kube-api-access-lmkkk\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: E1010 09:13:30.986014 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 09:13:31.485990119 +0000 UTC m=+154.502008861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.986438 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-catalog-content\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.986745 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-utilities\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.989978 4669 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-10T09:13:30.518442463Z","Handler":null,"Name":""} Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.997166 4669 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 10 09:13:30 crc kubenswrapper[4669]: I1010 09:13:30.997211 4669 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.009822 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmkkk\" (UniqueName: \"kubernetes.io/projected/cf42d4f9-7d35-43d2-a115-c82b5386be00-kube-api-access-lmkkk\") pod \"redhat-operators-p9txp\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.054232 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.086826 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.089011 4669 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.089043 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.115950 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j9f2l"] Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.117080 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.137443 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9f2l"] Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.143409 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w8ngm\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.188382 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.188667 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qld4b\" (UniqueName: \"kubernetes.io/projected/0f6469de-e0b2-40a3-aaf9-514f951e48d9-kube-api-access-qld4b\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.188694 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-utilities\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.188907 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-catalog-content\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.290812 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qld4b\" (UniqueName: \"kubernetes.io/projected/0f6469de-e0b2-40a3-aaf9-514f951e48d9-kube-api-access-qld4b\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.290858 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-utilities\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.290886 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-catalog-content\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.291549 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-catalog-content\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.291717 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-utilities\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.334578 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qld4b\" (UniqueName: \"kubernetes.io/projected/0f6469de-e0b2-40a3-aaf9-514f951e48d9-kube-api-access-qld4b\") pod \"redhat-operators-j9f2l\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.339216 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.366435 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.433053 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.536222 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf327e89-3302-463b-aaaf-87391e96aae1" containerID="474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897" exitCode=0 Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.536438 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwjd4" event={"ID":"bf327e89-3302-463b-aaaf-87391e96aae1","Type":"ContainerDied","Data":"474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897"} Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.536499 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwjd4" event={"ID":"bf327e89-3302-463b-aaaf-87391e96aae1","Type":"ContainerStarted","Data":"29ea1940fe229217a8795584cc29c8ab05d40584e746a11d5ad04fd54350509a"} Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.577333 4669 generic.go:334] "Generic (PLEG): container finished" podID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerID="945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247" exitCode=0 Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.578365 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26c6v" event={"ID":"434cae8a-10fb-41c0-af9d-4756b10fc8cf","Type":"ContainerDied","Data":"945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247"} Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.578388 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26c6v" event={"ID":"434cae8a-10fb-41c0-af9d-4756b10fc8cf","Type":"ContainerStarted","Data":"2e258d674891486dbb132fee63fb1d131a762c82f5cefc044e898f38182f86f5"} Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.711094 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p9txp"] Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.821931 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.845406 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w8ngm"] Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.927400 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.936722 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.940330 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.940563 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 10 09:13:31 crc kubenswrapper[4669]: I1010 09:13:31.943562 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.000290 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.000361 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.102752 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.102890 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.103242 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.104323 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9f2l"] Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.126381 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.278104 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.631181 4669 generic.go:334] "Generic (PLEG): container finished" podID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerID="33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0" exitCode=0 Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.631411 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9f2l" event={"ID":"0f6469de-e0b2-40a3-aaf9-514f951e48d9","Type":"ContainerDied","Data":"33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0"} Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.631457 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9f2l" event={"ID":"0f6469de-e0b2-40a3-aaf9-514f951e48d9","Type":"ContainerStarted","Data":"a9edbdc93eb016f04fe241b009a2400dd04e2d957c734b90745a8a805ff70327"} Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.694898 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" event={"ID":"bfb7109a-2f1c-49c9-960b-0be9a7028b35","Type":"ContainerStarted","Data":"2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660"} Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.694949 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" event={"ID":"bfb7109a-2f1c-49c9-960b-0be9a7028b35","Type":"ContainerStarted","Data":"11cb1eb9a2d49fcf5a76649bf6cb74d47fc4b06ae2819a2457d1019bc4b6eed2"} Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.695273 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.713640 4669 generic.go:334] "Generic (PLEG): container finished" podID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerID="84644e668df20f5d9ed2a757f6f223551d9ecbdf14d8a5a41a927a8784d7b199" exitCode=0 Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.713745 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9txp" event={"ID":"cf42d4f9-7d35-43d2-a115-c82b5386be00","Type":"ContainerDied","Data":"84644e668df20f5d9ed2a757f6f223551d9ecbdf14d8a5a41a927a8784d7b199"} Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.714091 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9txp" event={"ID":"cf42d4f9-7d35-43d2-a115-c82b5386be00","Type":"ContainerStarted","Data":"c848ba9718e4ac4209863a57d844de8f3fbf62c7e89758ddf353653bf6e33b6e"} Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.733943 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" podStartSLOduration=133.733921253 podStartE2EDuration="2m13.733921253s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:32.728236189 +0000 UTC m=+155.744254931" watchObservedRunningTime="2025-10-10 09:13:32.733921253 +0000 UTC m=+155.749939985" Oct 10 09:13:32 crc kubenswrapper[4669]: I1010 09:13:32.791338 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 09:13:33 crc kubenswrapper[4669]: I1010 09:13:33.753079 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0c51b918-ecd0-4c5a-b4ec-173060a8d913","Type":"ContainerStarted","Data":"1d5ebdfb78c3478a8cc8dc3abbd05af3dc32143ac17c172a55f6e4dc8b037894"} Oct 10 09:13:34 crc kubenswrapper[4669]: I1010 09:13:34.899042 4669 generic.go:334] "Generic (PLEG): container finished" podID="0c51b918-ecd0-4c5a-b4ec-173060a8d913" containerID="947f56e077ec53e01c4bf83c844d9f38cd023687b1e50568a78a9734f79700af" exitCode=0 Oct 10 09:13:34 crc kubenswrapper[4669]: I1010 09:13:34.899276 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0c51b918-ecd0-4c5a-b4ec-173060a8d913","Type":"ContainerDied","Data":"947f56e077ec53e01c4bf83c844d9f38cd023687b1e50568a78a9734f79700af"} Oct 10 09:13:35 crc kubenswrapper[4669]: I1010 09:13:35.049766 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rnpbk" Oct 10 09:13:35 crc kubenswrapper[4669]: I1010 09:13:35.879811 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 09:13:35 crc kubenswrapper[4669]: I1010 09:13:35.880697 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:35 crc kubenswrapper[4669]: I1010 09:13:35.889612 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 10 09:13:35 crc kubenswrapper[4669]: I1010 09:13:35.891655 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 10 09:13:35 crc kubenswrapper[4669]: I1010 09:13:35.907174 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.000675 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.000800 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.103068 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.103148 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.103288 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.134667 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.257970 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.758886 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.835984 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kube-api-access\") pod \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.836047 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kubelet-dir\") pod \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\" (UID: \"0c51b918-ecd0-4c5a-b4ec-173060a8d913\") " Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.836410 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0c51b918-ecd0-4c5a-b4ec-173060a8d913" (UID: "0c51b918-ecd0-4c5a-b4ec-173060a8d913"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.842484 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0c51b918-ecd0-4c5a-b4ec-173060a8d913" (UID: "0c51b918-ecd0-4c5a-b4ec-173060a8d913"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.938705 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 09:13:36 crc kubenswrapper[4669]: I1010 09:13:36.938730 4669 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0c51b918-ecd0-4c5a-b4ec-173060a8d913-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 10 09:13:37 crc kubenswrapper[4669]: I1010 09:13:37.089920 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0c51b918-ecd0-4c5a-b4ec-173060a8d913","Type":"ContainerDied","Data":"1d5ebdfb78c3478a8cc8dc3abbd05af3dc32143ac17c172a55f6e4dc8b037894"} Oct 10 09:13:37 crc kubenswrapper[4669]: I1010 09:13:37.089960 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d5ebdfb78c3478a8cc8dc3abbd05af3dc32143ac17c172a55f6e4dc8b037894" Oct 10 09:13:37 crc kubenswrapper[4669]: I1010 09:13:37.090028 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 09:13:37 crc kubenswrapper[4669]: I1010 09:13:37.090653 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 09:13:37 crc kubenswrapper[4669]: I1010 09:13:37.678063 4669 patch_prober.go:28] interesting pod/console-f9d7485db-wqc7t container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 10 09:13:37 crc kubenswrapper[4669]: I1010 09:13:37.678379 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wqc7t" podUID="69730259-7614-4869-bc84-10e17d8d13df" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 10 09:13:38 crc kubenswrapper[4669]: I1010 09:13:38.210925 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab5efa60-0426-4bd9-bb82-5f90af4e9563","Type":"ContainerStarted","Data":"1bf205ec7cf06bd5863240df9ea7e941aeaf41cf940466612ecd55f9f318301d"} Oct 10 09:13:38 crc kubenswrapper[4669]: I1010 09:13:38.228334 4669 generic.go:334] "Generic (PLEG): container finished" podID="98efe68d-5663-488a-92ca-c02a288b524f" containerID="f8edf14c55d21fe2700b9025389717cb6da78032ec4ae2098d6d03ed8542851b" exitCode=0 Oct 10 09:13:38 crc kubenswrapper[4669]: I1010 09:13:38.228382 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" event={"ID":"98efe68d-5663-488a-92ca-c02a288b524f","Type":"ContainerDied","Data":"f8edf14c55d21fe2700b9025389717cb6da78032ec4ae2098d6d03ed8542851b"} Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.041064 4669 patch_prober.go:28] interesting pod/downloads-7954f5f757-ngvr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.041871 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ngvr9" podUID="3f266776-152e-4b8c-b4e4-98a578b5f238" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.041064 4669 patch_prober.go:28] interesting pod/downloads-7954f5f757-ngvr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.042210 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ngvr9" podUID="3f266776-152e-4b8c-b4e4-98a578b5f238" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.253229 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab5efa60-0426-4bd9-bb82-5f90af4e9563","Type":"ContainerStarted","Data":"8b9a1a2cd8a21eb723593e9ca539e602d5d66ba61149d11dbc6d8eb381c04443"} Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.693177 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.716483 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.716458458 podStartE2EDuration="4.716458458s" podCreationTimestamp="2025-10-10 09:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:13:39.270354205 +0000 UTC m=+162.286372967" watchObservedRunningTime="2025-10-10 09:13:39.716458458 +0000 UTC m=+162.732477220" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.807250 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efe68d-5663-488a-92ca-c02a288b524f-secret-volume\") pod \"98efe68d-5663-488a-92ca-c02a288b524f\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.807339 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jffg9\" (UniqueName: \"kubernetes.io/projected/98efe68d-5663-488a-92ca-c02a288b524f-kube-api-access-jffg9\") pod \"98efe68d-5663-488a-92ca-c02a288b524f\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.807390 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efe68d-5663-488a-92ca-c02a288b524f-config-volume\") pod \"98efe68d-5663-488a-92ca-c02a288b524f\" (UID: \"98efe68d-5663-488a-92ca-c02a288b524f\") " Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.808342 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98efe68d-5663-488a-92ca-c02a288b524f-config-volume" (OuterVolumeSpecName: "config-volume") pod "98efe68d-5663-488a-92ca-c02a288b524f" (UID: "98efe68d-5663-488a-92ca-c02a288b524f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.821046 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98efe68d-5663-488a-92ca-c02a288b524f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "98efe68d-5663-488a-92ca-c02a288b524f" (UID: "98efe68d-5663-488a-92ca-c02a288b524f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.832413 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98efe68d-5663-488a-92ca-c02a288b524f-kube-api-access-jffg9" (OuterVolumeSpecName: "kube-api-access-jffg9") pod "98efe68d-5663-488a-92ca-c02a288b524f" (UID: "98efe68d-5663-488a-92ca-c02a288b524f"). InnerVolumeSpecName "kube-api-access-jffg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.908704 4669 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98efe68d-5663-488a-92ca-c02a288b524f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.908747 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jffg9\" (UniqueName: \"kubernetes.io/projected/98efe68d-5663-488a-92ca-c02a288b524f-kube-api-access-jffg9\") on node \"crc\" DevicePath \"\"" Oct 10 09:13:39 crc kubenswrapper[4669]: I1010 09:13:39.908761 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98efe68d-5663-488a-92ca-c02a288b524f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:13:40 crc kubenswrapper[4669]: I1010 09:13:40.312466 4669 generic.go:334] "Generic (PLEG): container finished" podID="ab5efa60-0426-4bd9-bb82-5f90af4e9563" containerID="8b9a1a2cd8a21eb723593e9ca539e602d5d66ba61149d11dbc6d8eb381c04443" exitCode=0 Oct 10 09:13:40 crc kubenswrapper[4669]: I1010 09:13:40.312535 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab5efa60-0426-4bd9-bb82-5f90af4e9563","Type":"ContainerDied","Data":"8b9a1a2cd8a21eb723593e9ca539e602d5d66ba61149d11dbc6d8eb381c04443"} Oct 10 09:13:40 crc kubenswrapper[4669]: I1010 09:13:40.332698 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" event={"ID":"98efe68d-5663-488a-92ca-c02a288b524f","Type":"ContainerDied","Data":"c55ca4c5ebf485127a24b74d1c9420306b46f7d4ba625b42fe2eb58ec6ff26db"} Oct 10 09:13:40 crc kubenswrapper[4669]: I1010 09:13:40.332764 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79" Oct 10 09:13:40 crc kubenswrapper[4669]: I1010 09:13:40.332785 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55ca4c5ebf485127a24b74d1c9420306b46f7d4ba625b42fe2eb58ec6ff26db" Oct 10 09:13:41 crc kubenswrapper[4669]: I1010 09:13:41.246805 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:41 crc kubenswrapper[4669]: I1010 09:13:41.251306 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/20d5735d-7ca2-4824-9b5d-4bb39502a3dc-metrics-certs\") pod \"network-metrics-daemon-bhn82\" (UID: \"20d5735d-7ca2-4824-9b5d-4bb39502a3dc\") " pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:41 crc kubenswrapper[4669]: I1010 09:13:41.520033 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bhn82" Oct 10 09:13:47 crc kubenswrapper[4669]: I1010 09:13:47.680031 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:47 crc kubenswrapper[4669]: I1010 09:13:47.693038 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:13:49 crc kubenswrapper[4669]: I1010 09:13:49.048313 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ngvr9" Oct 10 09:13:51 crc kubenswrapper[4669]: I1010 09:13:51.345919 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:13:54 crc kubenswrapper[4669]: I1010 09:13:54.275227 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:13:54 crc kubenswrapper[4669]: I1010 09:13:54.275318 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:13:55 crc kubenswrapper[4669]: I1010 09:13:55.751664 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:55 crc kubenswrapper[4669]: I1010 09:13:55.854982 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kube-api-access\") pod \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " Oct 10 09:13:55 crc kubenswrapper[4669]: I1010 09:13:55.855137 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kubelet-dir\") pod \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\" (UID: \"ab5efa60-0426-4bd9-bb82-5f90af4e9563\") " Oct 10 09:13:55 crc kubenswrapper[4669]: I1010 09:13:55.855272 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ab5efa60-0426-4bd9-bb82-5f90af4e9563" (UID: "ab5efa60-0426-4bd9-bb82-5f90af4e9563"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:13:55 crc kubenswrapper[4669]: I1010 09:13:55.855571 4669 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 10 09:13:55 crc kubenswrapper[4669]: I1010 09:13:55.860529 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ab5efa60-0426-4bd9-bb82-5f90af4e9563" (UID: "ab5efa60-0426-4bd9-bb82-5f90af4e9563"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:13:55 crc kubenswrapper[4669]: I1010 09:13:55.957561 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab5efa60-0426-4bd9-bb82-5f90af4e9563-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 09:13:56 crc kubenswrapper[4669]: I1010 09:13:56.468565 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab5efa60-0426-4bd9-bb82-5f90af4e9563","Type":"ContainerDied","Data":"1bf205ec7cf06bd5863240df9ea7e941aeaf41cf940466612ecd55f9f318301d"} Oct 10 09:13:56 crc kubenswrapper[4669]: I1010 09:13:56.468930 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bf205ec7cf06bd5863240df9ea7e941aeaf41cf940466612ecd55f9f318301d" Oct 10 09:13:56 crc kubenswrapper[4669]: I1010 09:13:56.469086 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 09:13:59 crc kubenswrapper[4669]: I1010 09:13:59.254849 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bngrg" Oct 10 09:14:01 crc kubenswrapper[4669]: I1010 09:14:01.814656 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-bhn82"] Oct 10 09:14:04 crc kubenswrapper[4669]: I1010 09:14:04.539767 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bhn82" event={"ID":"20d5735d-7ca2-4824-9b5d-4bb39502a3dc","Type":"ContainerStarted","Data":"86cd2b7ed03dde4f9bb38cc4bf412cfa0b942cfc294df7dd75dedd16a94382a2"} Oct 10 09:14:06 crc kubenswrapper[4669]: I1010 09:14:06.039946 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 09:14:07 crc kubenswrapper[4669]: E1010 09:14:07.513984 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 10 09:14:07 crc kubenswrapper[4669]: E1010 09:14:07.514494 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qld4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-j9f2l_openshift-marketplace(0f6469de-e0b2-40a3-aaf9-514f951e48d9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:07 crc kubenswrapper[4669]: E1010 09:14:07.515670 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-j9f2l" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" Oct 10 09:14:08 crc kubenswrapper[4669]: E1010 09:14:08.765429 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-j9f2l" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" Oct 10 09:14:08 crc kubenswrapper[4669]: E1010 09:14:08.841176 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 10 09:14:08 crc kubenswrapper[4669]: E1010 09:14:08.841370 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8b56k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cswwd_openshift-marketplace(efc3a420-b563-4f1b-8d25-17d363007b8d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:08 crc kubenswrapper[4669]: E1010 09:14:08.842619 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cswwd" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" Oct 10 09:14:08 crc kubenswrapper[4669]: E1010 09:14:08.884438 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 10 09:14:08 crc kubenswrapper[4669]: E1010 09:14:08.884721 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lmkkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-p9txp_openshift-marketplace(cf42d4f9-7d35-43d2-a115-c82b5386be00): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:08 crc kubenswrapper[4669]: E1010 09:14:08.885879 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-p9txp" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" Oct 10 09:14:09 crc kubenswrapper[4669]: E1010 09:14:09.522163 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 10 09:14:09 crc kubenswrapper[4669]: E1010 09:14:09.522340 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7h2fr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-26c6v_openshift-marketplace(434cae8a-10fb-41c0-af9d-4756b10fc8cf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:09 crc kubenswrapper[4669]: E1010 09:14:09.523508 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-26c6v" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.603918 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-p9txp" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.604070 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cswwd" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.604150 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-26c6v" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.809040 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.809514 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vsfmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-csltr_openshift-marketplace(85775a5b-bc26-43ba-a512-544217dec697): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.811568 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-csltr" podUID="85775a5b-bc26-43ba-a512-544217dec697" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.922875 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.923256 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dnnwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mwjd4_openshift-marketplace(bf327e89-3302-463b-aaaf-87391e96aae1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:11 crc kubenswrapper[4669]: E1010 09:14:11.931684 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mwjd4" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.034204 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.034363 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-762w7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d7z4x_openshift-marketplace(f38b0724-9f7b-4108-926f-e2192c26c49e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.035485 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-d7z4x" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.117650 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.117901 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7zwz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wlqpr_openshift-marketplace(42e85702-8412-4598-beb9-0e0ab5ae944e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.119057 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wlqpr" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" Oct 10 09:14:12 crc kubenswrapper[4669]: I1010 09:14:12.591101 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bhn82" event={"ID":"20d5735d-7ca2-4824-9b5d-4bb39502a3dc","Type":"ContainerStarted","Data":"b311f8cbcfd19b745768ea64a113b86e41bfa7d31872f3fcf428c6f99bfd1d3a"} Oct 10 09:14:12 crc kubenswrapper[4669]: I1010 09:14:12.591168 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bhn82" event={"ID":"20d5735d-7ca2-4824-9b5d-4bb39502a3dc","Type":"ContainerStarted","Data":"7ad5e4fd5cdc483b08fd552505ff0a5993ce3410b90e26b02872d49a8f99bae0"} Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.595892 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-csltr" podUID="85775a5b-bc26-43ba-a512-544217dec697" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.595926 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mwjd4" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.595938 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d7z4x" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" Oct 10 09:14:12 crc kubenswrapper[4669]: E1010 09:14:12.596018 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wlqpr" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" Oct 10 09:14:12 crc kubenswrapper[4669]: I1010 09:14:12.629955 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-bhn82" podStartSLOduration=173.629933426 podStartE2EDuration="2m53.629933426s" podCreationTimestamp="2025-10-10 09:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:14:12.628614724 +0000 UTC m=+195.644633496" watchObservedRunningTime="2025-10-10 09:14:12.629933426 +0000 UTC m=+195.645952188" Oct 10 09:14:24 crc kubenswrapper[4669]: I1010 09:14:24.275239 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:14:24 crc kubenswrapper[4669]: I1010 09:14:24.276777 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:14:24 crc kubenswrapper[4669]: I1010 09:14:24.276861 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:14:24 crc kubenswrapper[4669]: I1010 09:14:24.277417 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:14:24 crc kubenswrapper[4669]: I1010 09:14:24.277503 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c" gracePeriod=600 Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.679528 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9f2l" event={"ID":"0f6469de-e0b2-40a3-aaf9-514f951e48d9","Type":"ContainerStarted","Data":"0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621"} Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.681151 4669 generic.go:334] "Generic (PLEG): container finished" podID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerID="18f1a80478b62b58c8287a35cde1292f86b12eed393bffedb823f812b626e181" exitCode=0 Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.681193 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cswwd" event={"ID":"efc3a420-b563-4f1b-8d25-17d363007b8d","Type":"ContainerDied","Data":"18f1a80478b62b58c8287a35cde1292f86b12eed393bffedb823f812b626e181"} Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.686573 4669 generic.go:334] "Generic (PLEG): container finished" podID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerID="16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7" exitCode=0 Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.686692 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlqpr" event={"ID":"42e85702-8412-4598-beb9-0e0ab5ae944e","Type":"ContainerDied","Data":"16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7"} Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.689441 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9txp" event={"ID":"cf42d4f9-7d35-43d2-a115-c82b5386be00","Type":"ContainerStarted","Data":"0e99149308a39eeb84dd84c151f8e6fd7a509904d6b8ad92bea87967db54c301"} Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.693803 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csltr" event={"ID":"85775a5b-bc26-43ba-a512-544217dec697","Type":"ContainerStarted","Data":"1aa9d49220ea44358119d79d2df3394ce7f75388695dbe80dbd6b596190393d4"} Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.696387 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c" exitCode=0 Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.696432 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c"} Oct 10 09:14:25 crc kubenswrapper[4669]: I1010 09:14:25.696450 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"656cfe9a9a1a95c5e47506e7b135dc67d44e4e97a869b1e6e14c485ee25b4511"} Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.712037 4669 generic.go:334] "Generic (PLEG): container finished" podID="85775a5b-bc26-43ba-a512-544217dec697" containerID="1aa9d49220ea44358119d79d2df3394ce7f75388695dbe80dbd6b596190393d4" exitCode=0 Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.712162 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csltr" event={"ID":"85775a5b-bc26-43ba-a512-544217dec697","Type":"ContainerDied","Data":"1aa9d49220ea44358119d79d2df3394ce7f75388695dbe80dbd6b596190393d4"} Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.719226 4669 generic.go:334] "Generic (PLEG): container finished" podID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerID="f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f" exitCode=0 Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.719276 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26c6v" event={"ID":"434cae8a-10fb-41c0-af9d-4756b10fc8cf","Type":"ContainerDied","Data":"f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f"} Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.724556 4669 generic.go:334] "Generic (PLEG): container finished" podID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerID="0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621" exitCode=0 Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.724620 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9f2l" event={"ID":"0f6469de-e0b2-40a3-aaf9-514f951e48d9","Type":"ContainerDied","Data":"0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621"} Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.728286 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cswwd" event={"ID":"efc3a420-b563-4f1b-8d25-17d363007b8d","Type":"ContainerStarted","Data":"b5ba4cdc664f89fec9b62845c767e1b0931e7609f20f7434b3f0098ecc83e62a"} Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.732429 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlqpr" event={"ID":"42e85702-8412-4598-beb9-0e0ab5ae944e","Type":"ContainerStarted","Data":"04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93"} Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.735269 4669 generic.go:334] "Generic (PLEG): container finished" podID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerID="0e99149308a39eeb84dd84c151f8e6fd7a509904d6b8ad92bea87967db54c301" exitCode=0 Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.735295 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9txp" event={"ID":"cf42d4f9-7d35-43d2-a115-c82b5386be00","Type":"ContainerDied","Data":"0e99149308a39eeb84dd84c151f8e6fd7a509904d6b8ad92bea87967db54c301"} Oct 10 09:14:26 crc kubenswrapper[4669]: I1010 09:14:26.785469 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cswwd" podStartSLOduration=4.147020919 podStartE2EDuration="59.785455579s" podCreationTimestamp="2025-10-10 09:13:27 +0000 UTC" firstStartedPulling="2025-10-10 09:13:30.419725995 +0000 UTC m=+153.435744737" lastFinishedPulling="2025-10-10 09:14:26.058160655 +0000 UTC m=+209.074179397" observedRunningTime="2025-10-10 09:14:26.782636797 +0000 UTC m=+209.798655559" watchObservedRunningTime="2025-10-10 09:14:26.785455579 +0000 UTC m=+209.801474311" Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.742001 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf327e89-3302-463b-aaaf-87391e96aae1" containerID="b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d" exitCode=0 Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.742188 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwjd4" event={"ID":"bf327e89-3302-463b-aaaf-87391e96aae1","Type":"ContainerDied","Data":"b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d"} Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.745845 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26c6v" event={"ID":"434cae8a-10fb-41c0-af9d-4756b10fc8cf","Type":"ContainerStarted","Data":"ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51"} Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.748206 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9f2l" event={"ID":"0f6469de-e0b2-40a3-aaf9-514f951e48d9","Type":"ContainerStarted","Data":"8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886"} Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.750898 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9txp" event={"ID":"cf42d4f9-7d35-43d2-a115-c82b5386be00","Type":"ContainerStarted","Data":"e0d91957287c4c78151b7244497dc3d1048d3bd9d792e81bf1b773db3632a8e5"} Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.752769 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csltr" event={"ID":"85775a5b-bc26-43ba-a512-544217dec697","Type":"ContainerStarted","Data":"abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840"} Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.766921 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wlqpr" podStartSLOduration=4.933992027 podStartE2EDuration="1m0.766901852s" podCreationTimestamp="2025-10-10 09:13:27 +0000 UTC" firstStartedPulling="2025-10-10 09:13:30.434998528 +0000 UTC m=+153.451017270" lastFinishedPulling="2025-10-10 09:14:26.267908353 +0000 UTC m=+209.283927095" observedRunningTime="2025-10-10 09:14:26.867227079 +0000 UTC m=+209.883245821" watchObservedRunningTime="2025-10-10 09:14:27.766901852 +0000 UTC m=+210.782920594" Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.788258 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-26c6v" podStartSLOduration=3.209430791 podStartE2EDuration="58.788234968s" podCreationTimestamp="2025-10-10 09:13:29 +0000 UTC" firstStartedPulling="2025-10-10 09:13:31.58211753 +0000 UTC m=+154.598136272" lastFinishedPulling="2025-10-10 09:14:27.160921707 +0000 UTC m=+210.176940449" observedRunningTime="2025-10-10 09:14:27.785899682 +0000 UTC m=+210.801918424" watchObservedRunningTime="2025-10-10 09:14:27.788234968 +0000 UTC m=+210.804253710" Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.821471 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-csltr" podStartSLOduration=3.832222954 podStartE2EDuration="1m0.821451152s" podCreationTimestamp="2025-10-10 09:13:27 +0000 UTC" firstStartedPulling="2025-10-10 09:13:30.46910813 +0000 UTC m=+153.485126872" lastFinishedPulling="2025-10-10 09:14:27.458336328 +0000 UTC m=+210.474355070" observedRunningTime="2025-10-10 09:14:27.819817369 +0000 UTC m=+210.835836111" watchObservedRunningTime="2025-10-10 09:14:27.821451152 +0000 UTC m=+210.837469894" Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.847081 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p9txp" podStartSLOduration=3.353778513 podStartE2EDuration="57.847064528s" podCreationTimestamp="2025-10-10 09:13:30 +0000 UTC" firstStartedPulling="2025-10-10 09:13:32.723977522 +0000 UTC m=+155.739996264" lastFinishedPulling="2025-10-10 09:14:27.217263547 +0000 UTC m=+210.233282279" observedRunningTime="2025-10-10 09:14:27.843672108 +0000 UTC m=+210.859690850" watchObservedRunningTime="2025-10-10 09:14:27.847064528 +0000 UTC m=+210.863083270" Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.876399 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j9f2l" podStartSLOduration=2.139920108 podStartE2EDuration="56.876372505s" podCreationTimestamp="2025-10-10 09:13:31 +0000 UTC" firstStartedPulling="2025-10-10 09:13:32.645674992 +0000 UTC m=+155.661693734" lastFinishedPulling="2025-10-10 09:14:27.382127389 +0000 UTC m=+210.398146131" observedRunningTime="2025-10-10 09:14:27.873927725 +0000 UTC m=+210.889946467" watchObservedRunningTime="2025-10-10 09:14:27.876372505 +0000 UTC m=+210.892391247" Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.897996 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:14:27 crc kubenswrapper[4669]: I1010 09:14:27.898427 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:14:28 crc kubenswrapper[4669]: I1010 09:14:28.214347 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:14:28 crc kubenswrapper[4669]: I1010 09:14:28.214835 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:14:28 crc kubenswrapper[4669]: I1010 09:14:28.306736 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:14:28 crc kubenswrapper[4669]: I1010 09:14:28.306794 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:14:28 crc kubenswrapper[4669]: I1010 09:14:28.777340 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwjd4" event={"ID":"bf327e89-3302-463b-aaaf-87391e96aae1","Type":"ContainerStarted","Data":"331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e"} Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.039974 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-csltr" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="registry-server" probeResult="failure" output=< Oct 10 09:14:29 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:14:29 crc kubenswrapper[4669]: > Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.266167 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-cswwd" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="registry-server" probeResult="failure" output=< Oct 10 09:14:29 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:14:29 crc kubenswrapper[4669]: > Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.354227 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-wlqpr" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="registry-server" probeResult="failure" output=< Oct 10 09:14:29 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:14:29 crc kubenswrapper[4669]: > Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.787116 4669 generic.go:334] "Generic (PLEG): container finished" podID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerID="4a15ec8033e5e6f6971cc3cfcb615715d6d848f8b186e8d0dae1df9991a3f418" exitCode=0 Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.787167 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7z4x" event={"ID":"f38b0724-9f7b-4108-926f-e2192c26c49e","Type":"ContainerDied","Data":"4a15ec8033e5e6f6971cc3cfcb615715d6d848f8b186e8d0dae1df9991a3f418"} Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.819957 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mwjd4" podStartSLOduration=4.192005013 podStartE2EDuration="1m0.819937928s" podCreationTimestamp="2025-10-10 09:13:29 +0000 UTC" firstStartedPulling="2025-10-10 09:13:31.540805785 +0000 UTC m=+154.556824517" lastFinishedPulling="2025-10-10 09:14:28.16873869 +0000 UTC m=+211.184757432" observedRunningTime="2025-10-10 09:14:28.812783127 +0000 UTC m=+211.828801869" watchObservedRunningTime="2025-10-10 09:14:29.819937928 +0000 UTC m=+212.835956670" Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.890889 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.891138 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:14:29 crc kubenswrapper[4669]: I1010 09:14:29.969816 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:14:30 crc kubenswrapper[4669]: I1010 09:14:30.303102 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:14:30 crc kubenswrapper[4669]: I1010 09:14:30.303148 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:14:30 crc kubenswrapper[4669]: I1010 09:14:30.363826 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:14:30 crc kubenswrapper[4669]: I1010 09:14:30.798074 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7z4x" event={"ID":"f38b0724-9f7b-4108-926f-e2192c26c49e","Type":"ContainerStarted","Data":"ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff"} Oct 10 09:14:30 crc kubenswrapper[4669]: I1010 09:14:30.819798 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d7z4x" podStartSLOduration=3.983103515 podStartE2EDuration="1m3.819780091s" podCreationTimestamp="2025-10-10 09:13:27 +0000 UTC" firstStartedPulling="2025-10-10 09:13:30.454760866 +0000 UTC m=+153.470779598" lastFinishedPulling="2025-10-10 09:14:30.291437432 +0000 UTC m=+213.307456174" observedRunningTime="2025-10-10 09:14:30.817244188 +0000 UTC m=+213.833262930" watchObservedRunningTime="2025-10-10 09:14:30.819780091 +0000 UTC m=+213.835798833" Oct 10 09:14:31 crc kubenswrapper[4669]: I1010 09:14:31.055982 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:14:31 crc kubenswrapper[4669]: I1010 09:14:31.056041 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:14:31 crc kubenswrapper[4669]: I1010 09:14:31.433995 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:14:31 crc kubenswrapper[4669]: I1010 09:14:31.434040 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:14:32 crc kubenswrapper[4669]: I1010 09:14:32.103629 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p9txp" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="registry-server" probeResult="failure" output=< Oct 10 09:14:32 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:14:32 crc kubenswrapper[4669]: > Oct 10 09:14:32 crc kubenswrapper[4669]: I1010 09:14:32.474451 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j9f2l" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="registry-server" probeResult="failure" output=< Oct 10 09:14:32 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:14:32 crc kubenswrapper[4669]: > Oct 10 09:14:37 crc kubenswrapper[4669]: I1010 09:14:37.712627 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:14:37 crc kubenswrapper[4669]: I1010 09:14:37.713259 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:14:37 crc kubenswrapper[4669]: I1010 09:14:37.757423 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:14:37 crc kubenswrapper[4669]: I1010 09:14:37.868492 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:14:37 crc kubenswrapper[4669]: I1010 09:14:37.980187 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:14:38 crc kubenswrapper[4669]: I1010 09:14:38.016786 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:14:38 crc kubenswrapper[4669]: I1010 09:14:38.256274 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:14:38 crc kubenswrapper[4669]: I1010 09:14:38.295091 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:14:38 crc kubenswrapper[4669]: I1010 09:14:38.352886 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:14:38 crc kubenswrapper[4669]: I1010 09:14:38.394300 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:14:39 crc kubenswrapper[4669]: I1010 09:14:39.583955 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlqpr"] Oct 10 09:14:39 crc kubenswrapper[4669]: I1010 09:14:39.843655 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wlqpr" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="registry-server" containerID="cri-o://04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93" gracePeriod=2 Oct 10 09:14:39 crc kubenswrapper[4669]: I1010 09:14:39.927244 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:14:40 crc kubenswrapper[4669]: I1010 09:14:40.348068 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:14:40 crc kubenswrapper[4669]: I1010 09:14:40.583608 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cswwd"] Oct 10 09:14:40 crc kubenswrapper[4669]: I1010 09:14:40.583845 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cswwd" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="registry-server" containerID="cri-o://b5ba4cdc664f89fec9b62845c767e1b0931e7609f20f7434b3f0098ecc83e62a" gracePeriod=2 Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.093570 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.137348 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.484002 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.541717 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.610262 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.749173 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7zwz\" (UniqueName: \"kubernetes.io/projected/42e85702-8412-4598-beb9-0e0ab5ae944e-kube-api-access-w7zwz\") pod \"42e85702-8412-4598-beb9-0e0ab5ae944e\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.749301 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-catalog-content\") pod \"42e85702-8412-4598-beb9-0e0ab5ae944e\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.749389 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-utilities\") pod \"42e85702-8412-4598-beb9-0e0ab5ae944e\" (UID: \"42e85702-8412-4598-beb9-0e0ab5ae944e\") " Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.750389 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-utilities" (OuterVolumeSpecName: "utilities") pod "42e85702-8412-4598-beb9-0e0ab5ae944e" (UID: "42e85702-8412-4598-beb9-0e0ab5ae944e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.757479 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42e85702-8412-4598-beb9-0e0ab5ae944e-kube-api-access-w7zwz" (OuterVolumeSpecName: "kube-api-access-w7zwz") pod "42e85702-8412-4598-beb9-0e0ab5ae944e" (UID: "42e85702-8412-4598-beb9-0e0ab5ae944e"). InnerVolumeSpecName "kube-api-access-w7zwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.829474 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42e85702-8412-4598-beb9-0e0ab5ae944e" (UID: "42e85702-8412-4598-beb9-0e0ab5ae944e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.851488 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.851534 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e85702-8412-4598-beb9-0e0ab5ae944e-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.851549 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7zwz\" (UniqueName: \"kubernetes.io/projected/42e85702-8412-4598-beb9-0e0ab5ae944e-kube-api-access-w7zwz\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.859998 4669 generic.go:334] "Generic (PLEG): container finished" podID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerID="b5ba4cdc664f89fec9b62845c767e1b0931e7609f20f7434b3f0098ecc83e62a" exitCode=0 Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.860125 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cswwd" event={"ID":"efc3a420-b563-4f1b-8d25-17d363007b8d","Type":"ContainerDied","Data":"b5ba4cdc664f89fec9b62845c767e1b0931e7609f20f7434b3f0098ecc83e62a"} Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.862465 4669 generic.go:334] "Generic (PLEG): container finished" podID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerID="04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93" exitCode=0 Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.863113 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlqpr" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.863517 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlqpr" event={"ID":"42e85702-8412-4598-beb9-0e0ab5ae944e","Type":"ContainerDied","Data":"04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93"} Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.863574 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlqpr" event={"ID":"42e85702-8412-4598-beb9-0e0ab5ae944e","Type":"ContainerDied","Data":"7f74adaaec39cd3d56df61b81380455b8128ec4d9a4a0b41f7829781f99791df"} Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.863611 4669 scope.go:117] "RemoveContainer" containerID="04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.896403 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlqpr"] Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.897890 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wlqpr"] Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.899891 4669 scope.go:117] "RemoveContainer" containerID="16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.929896 4669 scope.go:117] "RemoveContainer" containerID="4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.970357 4669 scope.go:117] "RemoveContainer" containerID="04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93" Oct 10 09:14:41 crc kubenswrapper[4669]: E1010 09:14:41.970775 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93\": container with ID starting with 04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93 not found: ID does not exist" containerID="04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.970805 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93"} err="failed to get container status \"04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93\": rpc error: code = NotFound desc = could not find container \"04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93\": container with ID starting with 04e7103f01b53c1d2dae431f616e2c04f6b6c2237c65fb068532abb44ef05c93 not found: ID does not exist" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.970830 4669 scope.go:117] "RemoveContainer" containerID="16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7" Oct 10 09:14:41 crc kubenswrapper[4669]: E1010 09:14:41.971060 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7\": container with ID starting with 16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7 not found: ID does not exist" containerID="16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.971092 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7"} err="failed to get container status \"16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7\": rpc error: code = NotFound desc = could not find container \"16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7\": container with ID starting with 16a229a8187762915edaa42f50366c480b9efa09a2a99cb4106b4a87eda831f7 not found: ID does not exist" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.971107 4669 scope.go:117] "RemoveContainer" containerID="4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331" Oct 10 09:14:41 crc kubenswrapper[4669]: E1010 09:14:41.972310 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331\": container with ID starting with 4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331 not found: ID does not exist" containerID="4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.972329 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331"} err="failed to get container status \"4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331\": rpc error: code = NotFound desc = could not find container \"4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331\": container with ID starting with 4c805016cb3d50ded264f7cc7bd77a317bce24a1242b96c876074831a0fa7331 not found: ID does not exist" Oct 10 09:14:41 crc kubenswrapper[4669]: I1010 09:14:41.992284 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.052904 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-utilities\") pod \"efc3a420-b563-4f1b-8d25-17d363007b8d\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.053225 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b56k\" (UniqueName: \"kubernetes.io/projected/efc3a420-b563-4f1b-8d25-17d363007b8d-kube-api-access-8b56k\") pod \"efc3a420-b563-4f1b-8d25-17d363007b8d\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.053250 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-catalog-content\") pod \"efc3a420-b563-4f1b-8d25-17d363007b8d\" (UID: \"efc3a420-b563-4f1b-8d25-17d363007b8d\") " Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.053684 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-utilities" (OuterVolumeSpecName: "utilities") pod "efc3a420-b563-4f1b-8d25-17d363007b8d" (UID: "efc3a420-b563-4f1b-8d25-17d363007b8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.056921 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efc3a420-b563-4f1b-8d25-17d363007b8d-kube-api-access-8b56k" (OuterVolumeSpecName: "kube-api-access-8b56k") pod "efc3a420-b563-4f1b-8d25-17d363007b8d" (UID: "efc3a420-b563-4f1b-8d25-17d363007b8d"). InnerVolumeSpecName "kube-api-access-8b56k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.100295 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efc3a420-b563-4f1b-8d25-17d363007b8d" (UID: "efc3a420-b563-4f1b-8d25-17d363007b8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.155090 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.155121 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b56k\" (UniqueName: \"kubernetes.io/projected/efc3a420-b563-4f1b-8d25-17d363007b8d-kube-api-access-8b56k\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.155133 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efc3a420-b563-4f1b-8d25-17d363007b8d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.869161 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cswwd" event={"ID":"efc3a420-b563-4f1b-8d25-17d363007b8d","Type":"ContainerDied","Data":"dea38097e0a6b5751e80bc820cc7e63b674cc84b62b44dc5c3273bf0053e606b"} Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.869188 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cswwd" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.869239 4669 scope.go:117] "RemoveContainer" containerID="b5ba4cdc664f89fec9b62845c767e1b0931e7609f20f7434b3f0098ecc83e62a" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.884688 4669 scope.go:117] "RemoveContainer" containerID="18f1a80478b62b58c8287a35cde1292f86b12eed393bffedb823f812b626e181" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.899110 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cswwd"] Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.901808 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cswwd"] Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.915016 4669 scope.go:117] "RemoveContainer" containerID="b99a26a93642b2e3be49a6c4d75a910221e80209fa06c9048b352835d75ed127" Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.986001 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-26c6v"] Oct 10 09:14:42 crc kubenswrapper[4669]: I1010 09:14:42.986298 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-26c6v" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="registry-server" containerID="cri-o://ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51" gracePeriod=2 Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.285018 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.368776 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-catalog-content\") pod \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.369022 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-utilities\") pod \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.369125 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h2fr\" (UniqueName: \"kubernetes.io/projected/434cae8a-10fb-41c0-af9d-4756b10fc8cf-kube-api-access-7h2fr\") pod \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\" (UID: \"434cae8a-10fb-41c0-af9d-4756b10fc8cf\") " Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.372339 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-utilities" (OuterVolumeSpecName: "utilities") pod "434cae8a-10fb-41c0-af9d-4756b10fc8cf" (UID: "434cae8a-10fb-41c0-af9d-4756b10fc8cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.376434 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434cae8a-10fb-41c0-af9d-4756b10fc8cf-kube-api-access-7h2fr" (OuterVolumeSpecName: "kube-api-access-7h2fr") pod "434cae8a-10fb-41c0-af9d-4756b10fc8cf" (UID: "434cae8a-10fb-41c0-af9d-4756b10fc8cf"). InnerVolumeSpecName "kube-api-access-7h2fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.381413 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "434cae8a-10fb-41c0-af9d-4756b10fc8cf" (UID: "434cae8a-10fb-41c0-af9d-4756b10fc8cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.469952 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.469997 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h2fr\" (UniqueName: \"kubernetes.io/projected/434cae8a-10fb-41c0-af9d-4756b10fc8cf-kube-api-access-7h2fr\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.470008 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434cae8a-10fb-41c0-af9d-4756b10fc8cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.803291 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" path="/var/lib/kubelet/pods/42e85702-8412-4598-beb9-0e0ab5ae944e/volumes" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.803955 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" path="/var/lib/kubelet/pods/efc3a420-b563-4f1b-8d25-17d363007b8d/volumes" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.877674 4669 generic.go:334] "Generic (PLEG): container finished" podID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerID="ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51" exitCode=0 Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.877715 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26c6v" event={"ID":"434cae8a-10fb-41c0-af9d-4756b10fc8cf","Type":"ContainerDied","Data":"ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51"} Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.877741 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26c6v" event={"ID":"434cae8a-10fb-41c0-af9d-4756b10fc8cf","Type":"ContainerDied","Data":"2e258d674891486dbb132fee63fb1d131a762c82f5cefc044e898f38182f86f5"} Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.877757 4669 scope.go:117] "RemoveContainer" containerID="ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.877844 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26c6v" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.898519 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-26c6v"] Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.900240 4669 scope.go:117] "RemoveContainer" containerID="f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.901413 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-26c6v"] Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.922045 4669 scope.go:117] "RemoveContainer" containerID="945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.939632 4669 scope.go:117] "RemoveContainer" containerID="ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51" Oct 10 09:14:43 crc kubenswrapper[4669]: E1010 09:14:43.940282 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51\": container with ID starting with ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51 not found: ID does not exist" containerID="ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.940320 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51"} err="failed to get container status \"ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51\": rpc error: code = NotFound desc = could not find container \"ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51\": container with ID starting with ad8e0d469751238e2408bc23e6592a691b314776bc384e62f3f0ccdd83324d51 not found: ID does not exist" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.940346 4669 scope.go:117] "RemoveContainer" containerID="f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f" Oct 10 09:14:43 crc kubenswrapper[4669]: E1010 09:14:43.940790 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f\": container with ID starting with f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f not found: ID does not exist" containerID="f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.940830 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f"} err="failed to get container status \"f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f\": rpc error: code = NotFound desc = could not find container \"f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f\": container with ID starting with f7038732b00ff028cac7ebf3ffc088aae54d272832c5ddd17ba26b105ed91a9f not found: ID does not exist" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.940854 4669 scope.go:117] "RemoveContainer" containerID="945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247" Oct 10 09:14:43 crc kubenswrapper[4669]: E1010 09:14:43.941113 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247\": container with ID starting with 945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247 not found: ID does not exist" containerID="945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247" Oct 10 09:14:43 crc kubenswrapper[4669]: I1010 09:14:43.941133 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247"} err="failed to get container status \"945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247\": rpc error: code = NotFound desc = could not find container \"945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247\": container with ID starting with 945bac3a37bbe1100a07935934eb6194ff6fb0795ef61f426f4302cfb7523247 not found: ID does not exist" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.383644 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9f2l"] Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.383874 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j9f2l" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="registry-server" containerID="cri-o://8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886" gracePeriod=2 Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.721495 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.795514 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qld4b\" (UniqueName: \"kubernetes.io/projected/0f6469de-e0b2-40a3-aaf9-514f951e48d9-kube-api-access-qld4b\") pod \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.795565 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-catalog-content\") pod \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.795633 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-utilities\") pod \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\" (UID: \"0f6469de-e0b2-40a3-aaf9-514f951e48d9\") " Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.796449 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-utilities" (OuterVolumeSpecName: "utilities") pod "0f6469de-e0b2-40a3-aaf9-514f951e48d9" (UID: "0f6469de-e0b2-40a3-aaf9-514f951e48d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.800303 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f6469de-e0b2-40a3-aaf9-514f951e48d9-kube-api-access-qld4b" (OuterVolumeSpecName: "kube-api-access-qld4b") pod "0f6469de-e0b2-40a3-aaf9-514f951e48d9" (UID: "0f6469de-e0b2-40a3-aaf9-514f951e48d9"). InnerVolumeSpecName "kube-api-access-qld4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.802703 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" path="/var/lib/kubelet/pods/434cae8a-10fb-41c0-af9d-4756b10fc8cf/volumes" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.888413 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f6469de-e0b2-40a3-aaf9-514f951e48d9" (UID: "0f6469de-e0b2-40a3-aaf9-514f951e48d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.890798 4669 generic.go:334] "Generic (PLEG): container finished" podID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerID="8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886" exitCode=0 Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.890836 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9f2l" event={"ID":"0f6469de-e0b2-40a3-aaf9-514f951e48d9","Type":"ContainerDied","Data":"8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886"} Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.890863 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9f2l" event={"ID":"0f6469de-e0b2-40a3-aaf9-514f951e48d9","Type":"ContainerDied","Data":"a9edbdc93eb016f04fe241b009a2400dd04e2d957c734b90745a8a805ff70327"} Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.890891 4669 scope.go:117] "RemoveContainer" containerID="8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.891011 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9f2l" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.897120 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qld4b\" (UniqueName: \"kubernetes.io/projected/0f6469de-e0b2-40a3-aaf9-514f951e48d9-kube-api-access-qld4b\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.897165 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.897175 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6469de-e0b2-40a3-aaf9-514f951e48d9-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.922708 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9f2l"] Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.926453 4669 scope.go:117] "RemoveContainer" containerID="0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.927145 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j9f2l"] Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.943920 4669 scope.go:117] "RemoveContainer" containerID="33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.957305 4669 scope.go:117] "RemoveContainer" containerID="8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886" Oct 10 09:14:45 crc kubenswrapper[4669]: E1010 09:14:45.957642 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886\": container with ID starting with 8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886 not found: ID does not exist" containerID="8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.957678 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886"} err="failed to get container status \"8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886\": rpc error: code = NotFound desc = could not find container \"8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886\": container with ID starting with 8f7ecd68e64a9364f91b5a939deb3577cf5bba7541cdc9b7aa181188277ac886 not found: ID does not exist" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.957710 4669 scope.go:117] "RemoveContainer" containerID="0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621" Oct 10 09:14:45 crc kubenswrapper[4669]: E1010 09:14:45.958048 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621\": container with ID starting with 0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621 not found: ID does not exist" containerID="0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.958071 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621"} err="failed to get container status \"0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621\": rpc error: code = NotFound desc = could not find container \"0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621\": container with ID starting with 0e10be77fe05e0716f48bddfea2c49a9196e8e43ea695404d620505881900621 not found: ID does not exist" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.958086 4669 scope.go:117] "RemoveContainer" containerID="33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0" Oct 10 09:14:45 crc kubenswrapper[4669]: E1010 09:14:45.958250 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0\": container with ID starting with 33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0 not found: ID does not exist" containerID="33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0" Oct 10 09:14:45 crc kubenswrapper[4669]: I1010 09:14:45.958269 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0"} err="failed to get container status \"33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0\": rpc error: code = NotFound desc = could not find container \"33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0\": container with ID starting with 33dad947ae831c4b50de378d705c7e49aeab59e5d111fe3726d603b35a7086a0 not found: ID does not exist" Oct 10 09:14:47 crc kubenswrapper[4669]: I1010 09:14:47.805177 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" path="/var/lib/kubelet/pods/0f6469de-e0b2-40a3-aaf9-514f951e48d9/volumes" Oct 10 09:14:48 crc kubenswrapper[4669]: I1010 09:14:48.986286 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb9vp"] Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.143699 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf"] Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144429 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144444 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144458 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144465 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144475 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5efa60-0426-4bd9-bb82-5f90af4e9563" containerName="pruner" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144482 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5efa60-0426-4bd9-bb82-5f90af4e9563" containerName="pruner" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144496 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c51b918-ecd0-4c5a-b4ec-173060a8d913" containerName="pruner" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144502 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c51b918-ecd0-4c5a-b4ec-173060a8d913" containerName="pruner" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144511 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144517 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144528 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144535 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144545 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144551 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144561 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144568 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144579 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144601 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144615 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144623 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144634 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144641 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144651 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144658 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="extract-content" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144667 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144674 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144685 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144692 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="extract-utilities" Oct 10 09:15:00 crc kubenswrapper[4669]: E1010 09:15:00.144700 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98efe68d-5663-488a-92ca-c02a288b524f" containerName="collect-profiles" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144707 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="98efe68d-5663-488a-92ca-c02a288b524f" containerName="collect-profiles" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144817 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="434cae8a-10fb-41c0-af9d-4756b10fc8cf" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144828 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c51b918-ecd0-4c5a-b4ec-173060a8d913" containerName="pruner" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144851 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5efa60-0426-4bd9-bb82-5f90af4e9563" containerName="pruner" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144861 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f6469de-e0b2-40a3-aaf9-514f951e48d9" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144872 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc3a420-b563-4f1b-8d25-17d363007b8d" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144879 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="98efe68d-5663-488a-92ca-c02a288b524f" containerName="collect-profiles" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.144889 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e85702-8412-4598-beb9-0e0ab5ae944e" containerName="registry-server" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.147480 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.153538 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.154511 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.154898 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf"] Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.292462 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/973b9fdb-7b24-46a6-b351-25cbe8195797-config-volume\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.292568 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/973b9fdb-7b24-46a6-b351-25cbe8195797-secret-volume\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.292605 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrnhg\" (UniqueName: \"kubernetes.io/projected/973b9fdb-7b24-46a6-b351-25cbe8195797-kube-api-access-hrnhg\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.394559 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/973b9fdb-7b24-46a6-b351-25cbe8195797-config-volume\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.395763 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/973b9fdb-7b24-46a6-b351-25cbe8195797-config-volume\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.396030 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/973b9fdb-7b24-46a6-b351-25cbe8195797-secret-volume\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.396067 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrnhg\" (UniqueName: \"kubernetes.io/projected/973b9fdb-7b24-46a6-b351-25cbe8195797-kube-api-access-hrnhg\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.402017 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/973b9fdb-7b24-46a6-b351-25cbe8195797-secret-volume\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.420876 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrnhg\" (UniqueName: \"kubernetes.io/projected/973b9fdb-7b24-46a6-b351-25cbe8195797-kube-api-access-hrnhg\") pod \"collect-profiles-29334795-f2xqf\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.465398 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.666793 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf"] Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.981010 4669 generic.go:334] "Generic (PLEG): container finished" podID="973b9fdb-7b24-46a6-b351-25cbe8195797" containerID="49db09a03c7a401bd406b966d88aa62a27cf1a9e9625ca9d8bdbc064c7ac2f6a" exitCode=0 Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.981059 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" event={"ID":"973b9fdb-7b24-46a6-b351-25cbe8195797","Type":"ContainerDied","Data":"49db09a03c7a401bd406b966d88aa62a27cf1a9e9625ca9d8bdbc064c7ac2f6a"} Oct 10 09:15:00 crc kubenswrapper[4669]: I1010 09:15:00.981089 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" event={"ID":"973b9fdb-7b24-46a6-b351-25cbe8195797","Type":"ContainerStarted","Data":"ad81a6b0711e6996b94e214f5ef32616a558072a2b4904449f4562959c84f33c"} Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.197719 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.319760 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/973b9fdb-7b24-46a6-b351-25cbe8195797-config-volume\") pod \"973b9fdb-7b24-46a6-b351-25cbe8195797\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.319931 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrnhg\" (UniqueName: \"kubernetes.io/projected/973b9fdb-7b24-46a6-b351-25cbe8195797-kube-api-access-hrnhg\") pod \"973b9fdb-7b24-46a6-b351-25cbe8195797\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.319962 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/973b9fdb-7b24-46a6-b351-25cbe8195797-secret-volume\") pod \"973b9fdb-7b24-46a6-b351-25cbe8195797\" (UID: \"973b9fdb-7b24-46a6-b351-25cbe8195797\") " Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.320219 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/973b9fdb-7b24-46a6-b351-25cbe8195797-config-volume" (OuterVolumeSpecName: "config-volume") pod "973b9fdb-7b24-46a6-b351-25cbe8195797" (UID: "973b9fdb-7b24-46a6-b351-25cbe8195797"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.325332 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973b9fdb-7b24-46a6-b351-25cbe8195797-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "973b9fdb-7b24-46a6-b351-25cbe8195797" (UID: "973b9fdb-7b24-46a6-b351-25cbe8195797"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.325514 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973b9fdb-7b24-46a6-b351-25cbe8195797-kube-api-access-hrnhg" (OuterVolumeSpecName: "kube-api-access-hrnhg") pod "973b9fdb-7b24-46a6-b351-25cbe8195797" (UID: "973b9fdb-7b24-46a6-b351-25cbe8195797"). InnerVolumeSpecName "kube-api-access-hrnhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.421380 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrnhg\" (UniqueName: \"kubernetes.io/projected/973b9fdb-7b24-46a6-b351-25cbe8195797-kube-api-access-hrnhg\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.421423 4669 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/973b9fdb-7b24-46a6-b351-25cbe8195797-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.421432 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/973b9fdb-7b24-46a6-b351-25cbe8195797-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.992684 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" event={"ID":"973b9fdb-7b24-46a6-b351-25cbe8195797","Type":"ContainerDied","Data":"ad81a6b0711e6996b94e214f5ef32616a558072a2b4904449f4562959c84f33c"} Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.992740 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad81a6b0711e6996b94e214f5ef32616a558072a2b4904449f4562959c84f33c" Oct 10 09:15:02 crc kubenswrapper[4669]: I1010 09:15:02.992752 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.017068 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" podUID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" containerName="oauth-openshift" containerID="cri-o://fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2" gracePeriod=15 Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.413423 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.444128 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-fc667b7f-d88sk"] Oct 10 09:15:14 crc kubenswrapper[4669]: E1010 09:15:14.444561 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973b9fdb-7b24-46a6-b351-25cbe8195797" containerName="collect-profiles" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.444676 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="973b9fdb-7b24-46a6-b351-25cbe8195797" containerName="collect-profiles" Oct 10 09:15:14 crc kubenswrapper[4669]: E1010 09:15:14.444767 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" containerName="oauth-openshift" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.444846 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" containerName="oauth-openshift" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.445030 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" containerName="oauth-openshift" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.445125 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="973b9fdb-7b24-46a6-b351-25cbe8195797" containerName="collect-profiles" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.445661 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.459905 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fc667b7f-d88sk"] Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601094 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-login\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601193 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-idp-0-file-data\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601278 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-service-ca\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601356 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-error\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601422 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-trusted-ca-bundle\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601481 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-router-certs\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601529 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-session\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601641 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-provider-selection\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601765 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-serving-cert\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601831 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-dir\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601881 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-policies\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.601946 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-ocp-branding-template\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602007 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-584rk\" (UniqueName: \"kubernetes.io/projected/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-kube-api-access-584rk\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602061 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-cliconfig\") pod \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\" (UID: \"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4\") " Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602309 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602317 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602360 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602360 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602392 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-router-certs\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602449 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-error\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602523 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602646 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602713 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e7845de-c77f-4b8c-86dd-421cddfaab31-audit-dir\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602769 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.602840 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-session\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603157 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603171 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-service-ca\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603281 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603349 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpkr8\" (UniqueName: \"kubernetes.io/projected/9e7845de-c77f-4b8c-86dd-421cddfaab31-kube-api-access-fpkr8\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603403 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-login\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603427 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603561 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603610 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-audit-policies\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603754 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603772 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603786 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603801 4669 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.603814 4669 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.607276 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-kube-api-access-584rk" (OuterVolumeSpecName: "kube-api-access-584rk") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "kube-api-access-584rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.607757 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.608477 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.610063 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.610442 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.611466 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.611665 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.612403 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.613976 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" (UID: "105dbdd5-a3f4-46be-8bec-81bd45fbd7a4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.705061 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-login\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.705777 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.705827 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-audit-policies\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.705886 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.705929 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-router-certs\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.705964 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-error\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706008 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706054 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706089 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e7845de-c77f-4b8c-86dd-421cddfaab31-audit-dir\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706134 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706170 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-session\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706198 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-service-ca\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706225 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706262 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpkr8\" (UniqueName: \"kubernetes.io/projected/9e7845de-c77f-4b8c-86dd-421cddfaab31-kube-api-access-fpkr8\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706286 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9e7845de-c77f-4b8c-86dd-421cddfaab31-audit-dir\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706317 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706332 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706350 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706380 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706399 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706418 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706445 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-584rk\" (UniqueName: \"kubernetes.io/projected/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-kube-api-access-584rk\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706467 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.706496 4669 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.707056 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-audit-policies\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.708109 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.708262 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-service-ca\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.708565 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.709944 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-session\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.710764 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.713134 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-error\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.713838 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.713848 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-user-template-login\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.713983 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-router-certs\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.714237 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.714300 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e7845de-c77f-4b8c-86dd-421cddfaab31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.722300 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpkr8\" (UniqueName: \"kubernetes.io/projected/9e7845de-c77f-4b8c-86dd-421cddfaab31-kube-api-access-fpkr8\") pod \"oauth-openshift-fc667b7f-d88sk\" (UID: \"9e7845de-c77f-4b8c-86dd-421cddfaab31\") " pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.766455 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:14 crc kubenswrapper[4669]: I1010 09:15:14.939697 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fc667b7f-d88sk"] Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.057866 4669 generic.go:334] "Generic (PLEG): container finished" podID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" containerID="fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2" exitCode=0 Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.057955 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" event={"ID":"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4","Type":"ContainerDied","Data":"fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2"} Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.057996 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.058253 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pb9vp" event={"ID":"105dbdd5-a3f4-46be-8bec-81bd45fbd7a4","Type":"ContainerDied","Data":"84bdc7a3323e421ad75902dae8354f3b9df2bbc4e2d342daa63dc0e93b1c1678"} Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.058285 4669 scope.go:117] "RemoveContainer" containerID="fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2" Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.061949 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" event={"ID":"9e7845de-c77f-4b8c-86dd-421cddfaab31","Type":"ContainerStarted","Data":"130a50e35df507c7174356f3f9752279bb368ef8f2bb5b1e685ca7c6e4cc1574"} Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.079049 4669 scope.go:117] "RemoveContainer" containerID="fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2" Oct 10 09:15:15 crc kubenswrapper[4669]: E1010 09:15:15.079726 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2\": container with ID starting with fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2 not found: ID does not exist" containerID="fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2" Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.079767 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2"} err="failed to get container status \"fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2\": rpc error: code = NotFound desc = could not find container \"fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2\": container with ID starting with fed2af417939e4c8106477c042dbcab7c8aaeaa7307d94154f37b26ab6f8c1b2 not found: ID does not exist" Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.108750 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb9vp"] Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.112263 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pb9vp"] Oct 10 09:15:15 crc kubenswrapper[4669]: I1010 09:15:15.803429 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="105dbdd5-a3f4-46be-8bec-81bd45fbd7a4" path="/var/lib/kubelet/pods/105dbdd5-a3f4-46be-8bec-81bd45fbd7a4/volumes" Oct 10 09:15:16 crc kubenswrapper[4669]: I1010 09:15:16.069056 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" event={"ID":"9e7845de-c77f-4b8c-86dd-421cddfaab31","Type":"ContainerStarted","Data":"759bf576b217cfb6f1e89651c7d018e478a0d0b5d2bd716af2d580b434b7b156"} Oct 10 09:15:16 crc kubenswrapper[4669]: I1010 09:15:16.070112 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:16 crc kubenswrapper[4669]: I1010 09:15:16.078024 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" Oct 10 09:15:16 crc kubenswrapper[4669]: I1010 09:15:16.092237 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-fc667b7f-d88sk" podStartSLOduration=27.092210909 podStartE2EDuration="27.092210909s" podCreationTimestamp="2025-10-10 09:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:15:16.091230197 +0000 UTC m=+259.107248939" watchObservedRunningTime="2025-10-10 09:15:16.092210909 +0000 UTC m=+259.108229671" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.233451 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csltr"] Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.234281 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-csltr" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="registry-server" containerID="cri-o://abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840" gracePeriod=30 Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.243115 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7z4x"] Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.243458 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d7z4x" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="registry-server" containerID="cri-o://ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff" gracePeriod=30 Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.258245 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nvfcj"] Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.258507 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" podUID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" containerName="marketplace-operator" containerID="cri-o://18a0d221e8a00be5e90902c7004cc94602474bd7816f6a2bc0293f668cc84699" gracePeriod=30 Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.271866 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwjd4"] Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.272383 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mwjd4" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="registry-server" containerID="cri-o://331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e" gracePeriod=30 Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.275112 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p9txp"] Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.275631 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p9txp" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="registry-server" containerID="cri-o://e0d91957287c4c78151b7244497dc3d1048d3bd9d792e81bf1b773db3632a8e5" gracePeriod=30 Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.281229 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wf49x"] Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.282642 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.291344 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wf49x"] Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.411093 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/90231823-44be-4740-bef0-c98a048f9b05-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.411280 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fctbl\" (UniqueName: \"kubernetes.io/projected/90231823-44be-4740-bef0-c98a048f9b05-kube-api-access-fctbl\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.411499 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90231823-44be-4740-bef0-c98a048f9b05-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.512392 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fctbl\" (UniqueName: \"kubernetes.io/projected/90231823-44be-4740-bef0-c98a048f9b05-kube-api-access-fctbl\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.512456 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90231823-44be-4740-bef0-c98a048f9b05-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.512501 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/90231823-44be-4740-bef0-c98a048f9b05-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.514141 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/90231823-44be-4740-bef0-c98a048f9b05-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.521213 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/90231823-44be-4740-bef0-c98a048f9b05-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.529385 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fctbl\" (UniqueName: \"kubernetes.io/projected/90231823-44be-4740-bef0-c98a048f9b05-kube-api-access-fctbl\") pod \"marketplace-operator-79b997595-wf49x\" (UID: \"90231823-44be-4740-bef0-c98a048f9b05\") " pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.624890 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.718037 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff is running failed: container process not found" containerID="ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.726152 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff is running failed: container process not found" containerID="ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.726482 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff is running failed: container process not found" containerID="ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.726520 4669 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-d7z4x" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="registry-server" Oct 10 09:15:47 crc kubenswrapper[4669]: I1010 09:15:47.825724 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wf49x"] Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.898643 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840 is running failed: container process not found" containerID="abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.899315 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840 is running failed: container process not found" containerID="abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.899701 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840 is running failed: container process not found" containerID="abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840" cmd=["grpc_health_probe","-addr=:50051"] Oct 10 09:15:47 crc kubenswrapper[4669]: E1010 09:15:47.899747 4669 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-csltr" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="registry-server" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.129968 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.221107 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-catalog-content\") pod \"bf327e89-3302-463b-aaaf-87391e96aae1\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.221150 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnnwv\" (UniqueName: \"kubernetes.io/projected/bf327e89-3302-463b-aaaf-87391e96aae1-kube-api-access-dnnwv\") pod \"bf327e89-3302-463b-aaaf-87391e96aae1\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.221218 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-utilities\") pod \"bf327e89-3302-463b-aaaf-87391e96aae1\" (UID: \"bf327e89-3302-463b-aaaf-87391e96aae1\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.222709 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-utilities" (OuterVolumeSpecName: "utilities") pod "bf327e89-3302-463b-aaaf-87391e96aae1" (UID: "bf327e89-3302-463b-aaaf-87391e96aae1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.230037 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf327e89-3302-463b-aaaf-87391e96aae1-kube-api-access-dnnwv" (OuterVolumeSpecName: "kube-api-access-dnnwv") pod "bf327e89-3302-463b-aaaf-87391e96aae1" (UID: "bf327e89-3302-463b-aaaf-87391e96aae1"). InnerVolumeSpecName "kube-api-access-dnnwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.241835 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf327e89-3302-463b-aaaf-87391e96aae1" (UID: "bf327e89-3302-463b-aaaf-87391e96aae1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.276066 4669 generic.go:334] "Generic (PLEG): container finished" podID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerID="e0d91957287c4c78151b7244497dc3d1048d3bd9d792e81bf1b773db3632a8e5" exitCode=0 Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.276174 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9txp" event={"ID":"cf42d4f9-7d35-43d2-a115-c82b5386be00","Type":"ContainerDied","Data":"e0d91957287c4c78151b7244497dc3d1048d3bd9d792e81bf1b773db3632a8e5"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.281169 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.281602 4669 generic.go:334] "Generic (PLEG): container finished" podID="85775a5b-bc26-43ba-a512-544217dec697" containerID="abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840" exitCode=0 Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.281655 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csltr" event={"ID":"85775a5b-bc26-43ba-a512-544217dec697","Type":"ContainerDied","Data":"abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.283770 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf327e89-3302-463b-aaaf-87391e96aae1" containerID="331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e" exitCode=0 Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.283814 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwjd4" event={"ID":"bf327e89-3302-463b-aaaf-87391e96aae1","Type":"ContainerDied","Data":"331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.283831 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mwjd4" event={"ID":"bf327e89-3302-463b-aaaf-87391e96aae1","Type":"ContainerDied","Data":"29ea1940fe229217a8795584cc29c8ab05d40584e746a11d5ad04fd54350509a"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.283857 4669 scope.go:117] "RemoveContainer" containerID="331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.283973 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mwjd4" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.294639 4669 generic.go:334] "Generic (PLEG): container finished" podID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" containerID="18a0d221e8a00be5e90902c7004cc94602474bd7816f6a2bc0293f668cc84699" exitCode=0 Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.294900 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" event={"ID":"01b28f7e-7133-451a-b0d5-d1c0e13b3b40","Type":"ContainerDied","Data":"18a0d221e8a00be5e90902c7004cc94602474bd7816f6a2bc0293f668cc84699"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.295256 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nvfcj" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.307923 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" event={"ID":"90231823-44be-4740-bef0-c98a048f9b05","Type":"ContainerStarted","Data":"411736396bd8ab79ab152ec68fafe0f6a1919a2cfa9b70a1dc904249f6e6bdc9"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.308170 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" event={"ID":"90231823-44be-4740-bef0-c98a048f9b05","Type":"ContainerStarted","Data":"6b55f61b67a4fe7088d66cb9258dedc28279cc915c2e07bb65c9ce0bdebf89f7"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.310229 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.317219 4669 generic.go:334] "Generic (PLEG): container finished" podID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerID="ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff" exitCode=0 Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.317265 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7z4x" event={"ID":"f38b0724-9f7b-4108-926f-e2192c26c49e","Type":"ContainerDied","Data":"ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff"} Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.317361 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.321559 4669 scope.go:117] "RemoveContainer" containerID="b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.322116 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-trusted-ca\") pod \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.322148 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsfmq\" (UniqueName: \"kubernetes.io/projected/85775a5b-bc26-43ba-a512-544217dec697-kube-api-access-vsfmq\") pod \"85775a5b-bc26-43ba-a512-544217dec697\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.322265 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-operator-metrics\") pod \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.322291 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-utilities\") pod \"85775a5b-bc26-43ba-a512-544217dec697\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.322348 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmkkk\" (UniqueName: \"kubernetes.io/projected/cf42d4f9-7d35-43d2-a115-c82b5386be00-kube-api-access-lmkkk\") pod \"cf42d4f9-7d35-43d2-a115-c82b5386be00\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.322374 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-catalog-content\") pod \"cf42d4f9-7d35-43d2-a115-c82b5386be00\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.322408 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-utilities\") pod \"cf42d4f9-7d35-43d2-a115-c82b5386be00\" (UID: \"cf42d4f9-7d35-43d2-a115-c82b5386be00\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.323041 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ct8g\" (UniqueName: \"kubernetes.io/projected/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-kube-api-access-7ct8g\") pod \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\" (UID: \"01b28f7e-7133-451a-b0d5-d1c0e13b3b40\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.323125 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-catalog-content\") pod \"85775a5b-bc26-43ba-a512-544217dec697\" (UID: \"85775a5b-bc26-43ba-a512-544217dec697\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.323322 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.323339 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnnwv\" (UniqueName: \"kubernetes.io/projected/bf327e89-3302-463b-aaaf-87391e96aae1-kube-api-access-dnnwv\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.323352 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf327e89-3302-463b-aaaf-87391e96aae1-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.328027 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwjd4"] Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.329387 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-utilities" (OuterVolumeSpecName: "utilities") pod "85775a5b-bc26-43ba-a512-544217dec697" (UID: "85775a5b-bc26-43ba-a512-544217dec697"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.330131 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "01b28f7e-7133-451a-b0d5-d1c0e13b3b40" (UID: "01b28f7e-7133-451a-b0d5-d1c0e13b3b40"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.333174 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mwjd4"] Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.335459 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85775a5b-bc26-43ba-a512-544217dec697-kube-api-access-vsfmq" (OuterVolumeSpecName: "kube-api-access-vsfmq") pod "85775a5b-bc26-43ba-a512-544217dec697" (UID: "85775a5b-bc26-43ba-a512-544217dec697"). InnerVolumeSpecName "kube-api-access-vsfmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.337130 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "01b28f7e-7133-451a-b0d5-d1c0e13b3b40" (UID: "01b28f7e-7133-451a-b0d5-d1c0e13b3b40"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.337436 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-utilities" (OuterVolumeSpecName: "utilities") pod "cf42d4f9-7d35-43d2-a115-c82b5386be00" (UID: "cf42d4f9-7d35-43d2-a115-c82b5386be00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.341416 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf42d4f9-7d35-43d2-a115-c82b5386be00-kube-api-access-lmkkk" (OuterVolumeSpecName: "kube-api-access-lmkkk") pod "cf42d4f9-7d35-43d2-a115-c82b5386be00" (UID: "cf42d4f9-7d35-43d2-a115-c82b5386be00"). InnerVolumeSpecName "kube-api-access-lmkkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.347668 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.367039 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-kube-api-access-7ct8g" (OuterVolumeSpecName: "kube-api-access-7ct8g") pod "01b28f7e-7133-451a-b0d5-d1c0e13b3b40" (UID: "01b28f7e-7133-451a-b0d5-d1c0e13b3b40"). InnerVolumeSpecName "kube-api-access-7ct8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.383635 4669 scope.go:117] "RemoveContainer" containerID="474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.424492 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-utilities\") pod \"f38b0724-9f7b-4108-926f-e2192c26c49e\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.424629 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-762w7\" (UniqueName: \"kubernetes.io/projected/f38b0724-9f7b-4108-926f-e2192c26c49e-kube-api-access-762w7\") pod \"f38b0724-9f7b-4108-926f-e2192c26c49e\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.424661 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-catalog-content\") pod \"f38b0724-9f7b-4108-926f-e2192c26c49e\" (UID: \"f38b0724-9f7b-4108-926f-e2192c26c49e\") " Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.426054 4669 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.426120 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsfmq\" (UniqueName: \"kubernetes.io/projected/85775a5b-bc26-43ba-a512-544217dec697-kube-api-access-vsfmq\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.426132 4669 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.426144 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.426155 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmkkk\" (UniqueName: \"kubernetes.io/projected/cf42d4f9-7d35-43d2-a115-c82b5386be00-kube-api-access-lmkkk\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.426182 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.426194 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ct8g\" (UniqueName: \"kubernetes.io/projected/01b28f7e-7133-451a-b0d5-d1c0e13b3b40-kube-api-access-7ct8g\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.435132 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f38b0724-9f7b-4108-926f-e2192c26c49e-kube-api-access-762w7" (OuterVolumeSpecName: "kube-api-access-762w7") pod "f38b0724-9f7b-4108-926f-e2192c26c49e" (UID: "f38b0724-9f7b-4108-926f-e2192c26c49e"). InnerVolumeSpecName "kube-api-access-762w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.437141 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-utilities" (OuterVolumeSpecName: "utilities") pod "f38b0724-9f7b-4108-926f-e2192c26c49e" (UID: "f38b0724-9f7b-4108-926f-e2192c26c49e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.437864 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85775a5b-bc26-43ba-a512-544217dec697" (UID: "85775a5b-bc26-43ba-a512-544217dec697"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.452414 4669 scope.go:117] "RemoveContainer" containerID="331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e" Oct 10 09:15:48 crc kubenswrapper[4669]: E1010 09:15:48.456335 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e\": container with ID starting with 331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e not found: ID does not exist" containerID="331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.456391 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e"} err="failed to get container status \"331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e\": rpc error: code = NotFound desc = could not find container \"331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e\": container with ID starting with 331e253d77bf178c3045d492bb1c0fa9557d7bcc22c9a9f5f5467aa3deb2c01e not found: ID does not exist" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.456424 4669 scope.go:117] "RemoveContainer" containerID="b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d" Oct 10 09:15:48 crc kubenswrapper[4669]: E1010 09:15:48.458194 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d\": container with ID starting with b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d not found: ID does not exist" containerID="b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.458230 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d"} err="failed to get container status \"b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d\": rpc error: code = NotFound desc = could not find container \"b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d\": container with ID starting with b3871c4e61d12603b6e49ded413c7e9b93324cfa5bae000d520e8f01ed5c9a5d not found: ID does not exist" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.458249 4669 scope.go:117] "RemoveContainer" containerID="474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897" Oct 10 09:15:48 crc kubenswrapper[4669]: E1010 09:15:48.459332 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897\": container with ID starting with 474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897 not found: ID does not exist" containerID="474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.459357 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897"} err="failed to get container status \"474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897\": rpc error: code = NotFound desc = could not find container \"474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897\": container with ID starting with 474f18d059511834070b569bdd2ef7f79289e278b57b4ce6cdf4c5f9df005897 not found: ID does not exist" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.459370 4669 scope.go:117] "RemoveContainer" containerID="18a0d221e8a00be5e90902c7004cc94602474bd7816f6a2bc0293f668cc84699" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.477553 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f38b0724-9f7b-4108-926f-e2192c26c49e" (UID: "f38b0724-9f7b-4108-926f-e2192c26c49e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.508923 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf42d4f9-7d35-43d2-a115-c82b5386be00" (UID: "cf42d4f9-7d35-43d2-a115-c82b5386be00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.527170 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-762w7\" (UniqueName: \"kubernetes.io/projected/f38b0724-9f7b-4108-926f-e2192c26c49e-kube-api-access-762w7\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.527208 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.527218 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85775a5b-bc26-43ba-a512-544217dec697-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.527228 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf42d4f9-7d35-43d2-a115-c82b5386be00-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.527238 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f38b0724-9f7b-4108-926f-e2192c26c49e-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.621153 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nvfcj"] Oct 10 09:15:48 crc kubenswrapper[4669]: I1010 09:15:48.627430 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nvfcj"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.327134 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7z4x" event={"ID":"f38b0724-9f7b-4108-926f-e2192c26c49e","Type":"ContainerDied","Data":"9446e5e8490d001a695c2c6ae317ccaa097bd81da2c47bb3adf8d6178055b398"} Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.327196 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7z4x" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.327225 4669 scope.go:117] "RemoveContainer" containerID="ee673c1524a5f6df7096e6f89552419cd9bf962737fce7a09d67536439ad41ff" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.332116 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9txp" event={"ID":"cf42d4f9-7d35-43d2-a115-c82b5386be00","Type":"ContainerDied","Data":"c848ba9718e4ac4209863a57d844de8f3fbf62c7e89758ddf353653bf6e33b6e"} Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.332271 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p9txp" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.336657 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-csltr" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.337708 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-csltr" event={"ID":"85775a5b-bc26-43ba-a512-544217dec697","Type":"ContainerDied","Data":"8cc6f8647c0379240841e04538b4e7a1eee1a87c278fdcc4854afefea6573ea2"} Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.337784 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.341169 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.350855 4669 scope.go:117] "RemoveContainer" containerID="4a15ec8033e5e6f6971cc3cfcb615715d6d848f8b186e8d0dae1df9991a3f418" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.361748 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wf49x" podStartSLOduration=2.361731881 podStartE2EDuration="2.361731881s" podCreationTimestamp="2025-10-10 09:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:15:49.357286016 +0000 UTC m=+292.373304758" watchObservedRunningTime="2025-10-10 09:15:49.361731881 +0000 UTC m=+292.377750623" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.373249 4669 scope.go:117] "RemoveContainer" containerID="a2ff080160fbb8f27fc7f3a8dacd594a5ebed97ee77c3f4822bf7bf07785c424" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.390945 4669 scope.go:117] "RemoveContainer" containerID="e0d91957287c4c78151b7244497dc3d1048d3bd9d792e81bf1b773db3632a8e5" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.412701 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-csltr"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.430763 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-csltr"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.433880 4669 scope.go:117] "RemoveContainer" containerID="0e99149308a39eeb84dd84c151f8e6fd7a509904d6b8ad92bea87967db54c301" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.442518 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p9txp"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.448665 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p9txp"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.452645 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7z4x"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.455709 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d7z4x"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458274 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tltpl"] Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458493 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" containerName="marketplace-operator" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458516 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" containerName="marketplace-operator" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458560 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458570 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458596 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458602 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458611 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458619 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458629 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458636 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458653 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458660 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458675 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458682 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458692 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458698 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458706 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458712 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="extract-utilities" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458721 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458727 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458735 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458741 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458750 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458757 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: E1010 09:15:49.458764 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458770 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="extract-content" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458864 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458883 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458894 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="85775a5b-bc26-43ba-a512-544217dec697" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458904 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" containerName="marketplace-operator" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.458912 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" containerName="registry-server" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.459567 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.463225 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.463472 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tltpl"] Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.475964 4669 scope.go:117] "RemoveContainer" containerID="84644e668df20f5d9ed2a757f6f223551d9ecbdf14d8a5a41a927a8784d7b199" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.493689 4669 scope.go:117] "RemoveContainer" containerID="abeae4fe054720a0cf77e7df2d827acb5fd266ddaf0735ee270e30c2768f4840" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.509697 4669 scope.go:117] "RemoveContainer" containerID="1aa9d49220ea44358119d79d2df3394ce7f75388695dbe80dbd6b596190393d4" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.522429 4669 scope.go:117] "RemoveContainer" containerID="eba304693b0777d88678a70a67b6ee4dc7e0a865d39a43485e7b4c96f8980990" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.535398 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-catalog-content\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.535459 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgfcw\" (UniqueName: \"kubernetes.io/projected/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-kube-api-access-xgfcw\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.535499 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-utilities\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.636026 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-catalog-content\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.636111 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgfcw\" (UniqueName: \"kubernetes.io/projected/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-kube-api-access-xgfcw\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.636157 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-utilities\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.637257 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-catalog-content\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.638355 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-utilities\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.656569 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgfcw\" (UniqueName: \"kubernetes.io/projected/0aa2e009-01b6-46c5-bc27-41c9c62e46c5-kube-api-access-xgfcw\") pod \"redhat-marketplace-tltpl\" (UID: \"0aa2e009-01b6-46c5-bc27-41c9c62e46c5\") " pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.805965 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01b28f7e-7133-451a-b0d5-d1c0e13b3b40" path="/var/lib/kubelet/pods/01b28f7e-7133-451a-b0d5-d1c0e13b3b40/volumes" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.807257 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85775a5b-bc26-43ba-a512-544217dec697" path="/var/lib/kubelet/pods/85775a5b-bc26-43ba-a512-544217dec697/volumes" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.808669 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf327e89-3302-463b-aaaf-87391e96aae1" path="/var/lib/kubelet/pods/bf327e89-3302-463b-aaaf-87391e96aae1/volumes" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.809531 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.811014 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf42d4f9-7d35-43d2-a115-c82b5386be00" path="/var/lib/kubelet/pods/cf42d4f9-7d35-43d2-a115-c82b5386be00/volumes" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.811613 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f38b0724-9f7b-4108-926f-e2192c26c49e" path="/var/lib/kubelet/pods/f38b0724-9f7b-4108-926f-e2192c26c49e/volumes" Oct 10 09:15:49 crc kubenswrapper[4669]: I1010 09:15:49.987533 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tltpl"] Oct 10 09:15:49 crc kubenswrapper[4669]: W1010 09:15:49.990721 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0aa2e009_01b6_46c5_bc27_41c9c62e46c5.slice/crio-74bedc413b1e51beca288d8a66ff446c49e6e43f669cdf73671cd9c686ddef9f WatchSource:0}: Error finding container 74bedc413b1e51beca288d8a66ff446c49e6e43f669cdf73671cd9c686ddef9f: Status 404 returned error can't find the container with id 74bedc413b1e51beca288d8a66ff446c49e6e43f669cdf73671cd9c686ddef9f Oct 10 09:15:50 crc kubenswrapper[4669]: I1010 09:15:50.345337 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tltpl" event={"ID":"0aa2e009-01b6-46c5-bc27-41c9c62e46c5","Type":"ContainerStarted","Data":"74bedc413b1e51beca288d8a66ff446c49e6e43f669cdf73671cd9c686ddef9f"} Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.352666 4669 generic.go:334] "Generic (PLEG): container finished" podID="0aa2e009-01b6-46c5-bc27-41c9c62e46c5" containerID="3581b03ce4e3a76e8209741adbd6a0f12f7fbcd1dd9b28b92efdd1be8dbc442c" exitCode=0 Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.352785 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tltpl" event={"ID":"0aa2e009-01b6-46c5-bc27-41c9c62e46c5","Type":"ContainerDied","Data":"3581b03ce4e3a76e8209741adbd6a0f12f7fbcd1dd9b28b92efdd1be8dbc442c"} Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.642714 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fq9ss"] Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.644301 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.646877 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.651208 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fq9ss"] Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.760151 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a078913-33a1-429a-850e-400f184f3cc8-utilities\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.760440 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a078913-33a1-429a-850e-400f184f3cc8-catalog-content\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.760642 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vcct\" (UniqueName: \"kubernetes.io/projected/9a078913-33a1-429a-850e-400f184f3cc8-kube-api-access-9vcct\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.839698 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5b7lm"] Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.840975 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.843155 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.854761 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5b7lm"] Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.861946 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a078913-33a1-429a-850e-400f184f3cc8-utilities\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.861985 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a078913-33a1-429a-850e-400f184f3cc8-catalog-content\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.862056 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vcct\" (UniqueName: \"kubernetes.io/projected/9a078913-33a1-429a-850e-400f184f3cc8-kube-api-access-9vcct\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.862755 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a078913-33a1-429a-850e-400f184f3cc8-utilities\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.862974 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a078913-33a1-429a-850e-400f184f3cc8-catalog-content\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.913777 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vcct\" (UniqueName: \"kubernetes.io/projected/9a078913-33a1-429a-850e-400f184f3cc8-kube-api-access-9vcct\") pod \"community-operators-fq9ss\" (UID: \"9a078913-33a1-429a-850e-400f184f3cc8\") " pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.961427 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.963474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ff4b09-5116-45de-9f97-f87832000566-utilities\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.963549 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nttz\" (UniqueName: \"kubernetes.io/projected/e8ff4b09-5116-45de-9f97-f87832000566-kube-api-access-6nttz\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:51 crc kubenswrapper[4669]: I1010 09:15:51.963571 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ff4b09-5116-45de-9f97-f87832000566-catalog-content\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.064922 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ff4b09-5116-45de-9f97-f87832000566-utilities\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.064969 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nttz\" (UniqueName: \"kubernetes.io/projected/e8ff4b09-5116-45de-9f97-f87832000566-kube-api-access-6nttz\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.064990 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ff4b09-5116-45de-9f97-f87832000566-catalog-content\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.065690 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ff4b09-5116-45de-9f97-f87832000566-utilities\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.065724 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ff4b09-5116-45de-9f97-f87832000566-catalog-content\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.093863 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nttz\" (UniqueName: \"kubernetes.io/projected/e8ff4b09-5116-45de-9f97-f87832000566-kube-api-access-6nttz\") pod \"certified-operators-5b7lm\" (UID: \"e8ff4b09-5116-45de-9f97-f87832000566\") " pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.159882 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fq9ss"] Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.200753 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.366246 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fq9ss" event={"ID":"9a078913-33a1-429a-850e-400f184f3cc8","Type":"ContainerStarted","Data":"8bc00a579621c34cf288a6fb522e72157c6367f240a292f512b320c7cffe62e4"} Oct 10 09:15:52 crc kubenswrapper[4669]: I1010 09:15:52.375253 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5b7lm"] Oct 10 09:15:52 crc kubenswrapper[4669]: W1010 09:15:52.382322 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8ff4b09_5116_45de_9f97_f87832000566.slice/crio-4e659c5372234c81fc7883ee2a0a1e432c346e67228a3b030ac12eb0043c36b3 WatchSource:0}: Error finding container 4e659c5372234c81fc7883ee2a0a1e432c346e67228a3b030ac12eb0043c36b3: Status 404 returned error can't find the container with id 4e659c5372234c81fc7883ee2a0a1e432c346e67228a3b030ac12eb0043c36b3 Oct 10 09:15:53 crc kubenswrapper[4669]: I1010 09:15:53.372858 4669 generic.go:334] "Generic (PLEG): container finished" podID="e8ff4b09-5116-45de-9f97-f87832000566" containerID="a63da138da32b04de100b72d4ee9ba27a6ee09f80d6dff829509a3bc1b9b6cae" exitCode=0 Oct 10 09:15:53 crc kubenswrapper[4669]: I1010 09:15:53.373161 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b7lm" event={"ID":"e8ff4b09-5116-45de-9f97-f87832000566","Type":"ContainerDied","Data":"a63da138da32b04de100b72d4ee9ba27a6ee09f80d6dff829509a3bc1b9b6cae"} Oct 10 09:15:53 crc kubenswrapper[4669]: I1010 09:15:53.373189 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b7lm" event={"ID":"e8ff4b09-5116-45de-9f97-f87832000566","Type":"ContainerStarted","Data":"4e659c5372234c81fc7883ee2a0a1e432c346e67228a3b030ac12eb0043c36b3"} Oct 10 09:15:53 crc kubenswrapper[4669]: I1010 09:15:53.375089 4669 generic.go:334] "Generic (PLEG): container finished" podID="9a078913-33a1-429a-850e-400f184f3cc8" containerID="0d4009cbe06fa4226825c8a8ca61af14c3370ef65d7e9a5661df5b33ebccbd84" exitCode=0 Oct 10 09:15:53 crc kubenswrapper[4669]: I1010 09:15:53.375108 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fq9ss" event={"ID":"9a078913-33a1-429a-850e-400f184f3cc8","Type":"ContainerDied","Data":"0d4009cbe06fa4226825c8a8ca61af14c3370ef65d7e9a5661df5b33ebccbd84"} Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.043896 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-phbdn"] Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.045325 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.048367 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.055550 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phbdn"] Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.212156 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-utilities\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.212220 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-catalog-content\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.212239 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b5lz\" (UniqueName: \"kubernetes.io/projected/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-kube-api-access-4b5lz\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.323036 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-utilities\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.323132 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-catalog-content\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.323162 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b5lz\" (UniqueName: \"kubernetes.io/projected/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-kube-api-access-4b5lz\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.323646 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-utilities\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.323796 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-catalog-content\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.345391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b5lz\" (UniqueName: \"kubernetes.io/projected/1ce843ca-00a5-4c8a-abbb-dd9fb719f331-kube-api-access-4b5lz\") pod \"redhat-operators-phbdn\" (UID: \"1ce843ca-00a5-4c8a-abbb-dd9fb719f331\") " pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.360268 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.380604 4669 generic.go:334] "Generic (PLEG): container finished" podID="0aa2e009-01b6-46c5-bc27-41c9c62e46c5" containerID="954a290d0ef3c44b4640e7818ea2dbb663b8fa42a2ee3507db797b317c8af9ce" exitCode=0 Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.380807 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tltpl" event={"ID":"0aa2e009-01b6-46c5-bc27-41c9c62e46c5","Type":"ContainerDied","Data":"954a290d0ef3c44b4640e7818ea2dbb663b8fa42a2ee3507db797b317c8af9ce"} Oct 10 09:15:54 crc kubenswrapper[4669]: I1010 09:15:54.566770 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-phbdn"] Oct 10 09:15:55 crc kubenswrapper[4669]: I1010 09:15:55.387872 4669 generic.go:334] "Generic (PLEG): container finished" podID="1ce843ca-00a5-4c8a-abbb-dd9fb719f331" containerID="ff16a9e8b94c06839adaf1ad7beed29edde678a0810381ac9594487efbafc5de" exitCode=0 Oct 10 09:15:55 crc kubenswrapper[4669]: I1010 09:15:55.387981 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phbdn" event={"ID":"1ce843ca-00a5-4c8a-abbb-dd9fb719f331","Type":"ContainerDied","Data":"ff16a9e8b94c06839adaf1ad7beed29edde678a0810381ac9594487efbafc5de"} Oct 10 09:15:55 crc kubenswrapper[4669]: I1010 09:15:55.388148 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phbdn" event={"ID":"1ce843ca-00a5-4c8a-abbb-dd9fb719f331","Type":"ContainerStarted","Data":"5996fd5a9cf89e4d8486f4e33c4cce5c95444211a9195ac6476ac84b27132398"} Oct 10 09:15:58 crc kubenswrapper[4669]: I1010 09:15:58.405018 4669 generic.go:334] "Generic (PLEG): container finished" podID="9a078913-33a1-429a-850e-400f184f3cc8" containerID="84dcfc2d3884d4ba9fd69ef933c3e54f5ff8fd147333e2b9aae6fde4939e493f" exitCode=0 Oct 10 09:15:58 crc kubenswrapper[4669]: I1010 09:15:58.405118 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fq9ss" event={"ID":"9a078913-33a1-429a-850e-400f184f3cc8","Type":"ContainerDied","Data":"84dcfc2d3884d4ba9fd69ef933c3e54f5ff8fd147333e2b9aae6fde4939e493f"} Oct 10 09:15:59 crc kubenswrapper[4669]: I1010 09:15:59.413801 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tltpl" event={"ID":"0aa2e009-01b6-46c5-bc27-41c9c62e46c5","Type":"ContainerStarted","Data":"c19e24c7e6f225b7dc05962a54e62172941f399bdfef22910d641085b1580037"} Oct 10 09:15:59 crc kubenswrapper[4669]: I1010 09:15:59.416684 4669 generic.go:334] "Generic (PLEG): container finished" podID="e8ff4b09-5116-45de-9f97-f87832000566" containerID="c9e9c3db09a0af72f2771095f5ddd12d3f1fe88b6b052e8a5d365975f6877066" exitCode=0 Oct 10 09:15:59 crc kubenswrapper[4669]: I1010 09:15:59.416727 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b7lm" event={"ID":"e8ff4b09-5116-45de-9f97-f87832000566","Type":"ContainerDied","Data":"c9e9c3db09a0af72f2771095f5ddd12d3f1fe88b6b052e8a5d365975f6877066"} Oct 10 09:16:00 crc kubenswrapper[4669]: I1010 09:16:00.467041 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tltpl" podStartSLOduration=5.356780818 podStartE2EDuration="11.467019183s" podCreationTimestamp="2025-10-10 09:15:49 +0000 UTC" firstStartedPulling="2025-10-10 09:15:51.354245042 +0000 UTC m=+294.370263784" lastFinishedPulling="2025-10-10 09:15:57.464483407 +0000 UTC m=+300.480502149" observedRunningTime="2025-10-10 09:16:00.465267896 +0000 UTC m=+303.481286638" watchObservedRunningTime="2025-10-10 09:16:00.467019183 +0000 UTC m=+303.483037945" Oct 10 09:16:04 crc kubenswrapper[4669]: I1010 09:16:04.446857 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b7lm" event={"ID":"e8ff4b09-5116-45de-9f97-f87832000566","Type":"ContainerStarted","Data":"161396bb9ed4f3b23fd6494821899d2b0dd94a9cb613e566f2bb9ceef18f2933"} Oct 10 09:16:04 crc kubenswrapper[4669]: I1010 09:16:04.448788 4669 generic.go:334] "Generic (PLEG): container finished" podID="1ce843ca-00a5-4c8a-abbb-dd9fb719f331" containerID="5b4ea5f8927520480bb3b12d591c93c1fcc1c4031d7a9bac3c89a8581075dbbb" exitCode=0 Oct 10 09:16:04 crc kubenswrapper[4669]: I1010 09:16:04.448818 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phbdn" event={"ID":"1ce843ca-00a5-4c8a-abbb-dd9fb719f331","Type":"ContainerDied","Data":"5b4ea5f8927520480bb3b12d591c93c1fcc1c4031d7a9bac3c89a8581075dbbb"} Oct 10 09:16:04 crc kubenswrapper[4669]: I1010 09:16:04.465789 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5b7lm" podStartSLOduration=3.8682046039999998 podStartE2EDuration="13.465775703s" podCreationTimestamp="2025-10-10 09:15:51 +0000 UTC" firstStartedPulling="2025-10-10 09:15:53.554901408 +0000 UTC m=+296.570920150" lastFinishedPulling="2025-10-10 09:16:03.152472507 +0000 UTC m=+306.168491249" observedRunningTime="2025-10-10 09:16:04.46354436 +0000 UTC m=+307.479563102" watchObservedRunningTime="2025-10-10 09:16:04.465775703 +0000 UTC m=+307.481794445" Oct 10 09:16:06 crc kubenswrapper[4669]: I1010 09:16:06.465476 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fq9ss" event={"ID":"9a078913-33a1-429a-850e-400f184f3cc8","Type":"ContainerStarted","Data":"ab5214816283e89a06f63f671b7b0d0970d9b71161ba1f53546d9ade229fb4ec"} Oct 10 09:16:07 crc kubenswrapper[4669]: I1010 09:16:07.499696 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fq9ss" podStartSLOduration=5.624692109 podStartE2EDuration="16.499679443s" podCreationTimestamp="2025-10-10 09:15:51 +0000 UTC" firstStartedPulling="2025-10-10 09:15:53.387150041 +0000 UTC m=+296.403168783" lastFinishedPulling="2025-10-10 09:16:04.262137375 +0000 UTC m=+307.278156117" observedRunningTime="2025-10-10 09:16:07.499334631 +0000 UTC m=+310.515353393" watchObservedRunningTime="2025-10-10 09:16:07.499679443 +0000 UTC m=+310.515698195" Oct 10 09:16:09 crc kubenswrapper[4669]: I1010 09:16:09.810434 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:16:09 crc kubenswrapper[4669]: I1010 09:16:09.810494 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:16:09 crc kubenswrapper[4669]: I1010 09:16:09.866821 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:16:10 crc kubenswrapper[4669]: I1010 09:16:10.524762 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tltpl" Oct 10 09:16:11 crc kubenswrapper[4669]: I1010 09:16:11.965237 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:16:11 crc kubenswrapper[4669]: I1010 09:16:11.966149 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:16:12 crc kubenswrapper[4669]: I1010 09:16:12.011100 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:16:12 crc kubenswrapper[4669]: I1010 09:16:12.202144 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:16:12 crc kubenswrapper[4669]: I1010 09:16:12.202194 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:16:12 crc kubenswrapper[4669]: I1010 09:16:12.245339 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:16:12 crc kubenswrapper[4669]: I1010 09:16:12.539556 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fq9ss" Oct 10 09:16:12 crc kubenswrapper[4669]: I1010 09:16:12.545199 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5b7lm" Oct 10 09:16:16 crc kubenswrapper[4669]: I1010 09:16:16.529901 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-phbdn" event={"ID":"1ce843ca-00a5-4c8a-abbb-dd9fb719f331","Type":"ContainerStarted","Data":"2af4571d964d2cd139d4ada86b7a106e54dfbd647a2c034a9e193967936e06c1"} Oct 10 09:16:18 crc kubenswrapper[4669]: I1010 09:16:18.560948 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-phbdn" podStartSLOduration=5.8533116750000005 podStartE2EDuration="24.560930586s" podCreationTimestamp="2025-10-10 09:15:54 +0000 UTC" firstStartedPulling="2025-10-10 09:15:55.843163975 +0000 UTC m=+298.859182717" lastFinishedPulling="2025-10-10 09:16:14.550782876 +0000 UTC m=+317.566801628" observedRunningTime="2025-10-10 09:16:18.560010617 +0000 UTC m=+321.576029359" watchObservedRunningTime="2025-10-10 09:16:18.560930586 +0000 UTC m=+321.576949338" Oct 10 09:16:24 crc kubenswrapper[4669]: I1010 09:16:24.361125 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:16:24 crc kubenswrapper[4669]: I1010 09:16:24.361492 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:16:24 crc kubenswrapper[4669]: I1010 09:16:24.406088 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:16:24 crc kubenswrapper[4669]: I1010 09:16:24.629673 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-phbdn" Oct 10 09:16:54 crc kubenswrapper[4669]: I1010 09:16:54.274679 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:16:54 crc kubenswrapper[4669]: I1010 09:16:54.275517 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:17:24 crc kubenswrapper[4669]: I1010 09:17:24.274972 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:17:24 crc kubenswrapper[4669]: I1010 09:17:24.275647 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.307609 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-c7s9q"] Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.308475 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.365205 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-c7s9q"] Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400432 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk5c5\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-kube-api-access-rk5c5\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400493 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56453fbc-930c-4145-9104-ef8e74f89b85-ca-trust-extracted\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400517 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56453fbc-930c-4145-9104-ef8e74f89b85-trusted-ca\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400565 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56453fbc-930c-4145-9104-ef8e74f89b85-registry-certificates\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400614 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-registry-tls\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400646 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400663 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56453fbc-930c-4145-9104-ef8e74f89b85-installation-pull-secrets\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.400689 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-bound-sa-token\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.426555 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.501843 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-bound-sa-token\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.501899 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk5c5\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-kube-api-access-rk5c5\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.501951 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56453fbc-930c-4145-9104-ef8e74f89b85-ca-trust-extracted\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.501970 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56453fbc-930c-4145-9104-ef8e74f89b85-trusted-ca\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.502018 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56453fbc-930c-4145-9104-ef8e74f89b85-registry-certificates\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.502049 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-registry-tls\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.502080 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56453fbc-930c-4145-9104-ef8e74f89b85-installation-pull-secrets\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.502726 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56453fbc-930c-4145-9104-ef8e74f89b85-ca-trust-extracted\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.503454 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56453fbc-930c-4145-9104-ef8e74f89b85-registry-certificates\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.503861 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56453fbc-930c-4145-9104-ef8e74f89b85-trusted-ca\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.508183 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-registry-tls\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.508184 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56453fbc-930c-4145-9104-ef8e74f89b85-installation-pull-secrets\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.518414 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-bound-sa-token\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.520350 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk5c5\" (UniqueName: \"kubernetes.io/projected/56453fbc-930c-4145-9104-ef8e74f89b85-kube-api-access-rk5c5\") pod \"image-registry-66df7c8f76-c7s9q\" (UID: \"56453fbc-930c-4145-9104-ef8e74f89b85\") " pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.624304 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.805843 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-c7s9q"] Oct 10 09:17:31 crc kubenswrapper[4669]: W1010 09:17:31.812626 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56453fbc_930c_4145_9104_ef8e74f89b85.slice/crio-77d94d436927d52a68cecd7715348e068926c6500913de946185410e37f4ebd0 WatchSource:0}: Error finding container 77d94d436927d52a68cecd7715348e068926c6500913de946185410e37f4ebd0: Status 404 returned error can't find the container with id 77d94d436927d52a68cecd7715348e068926c6500913de946185410e37f4ebd0 Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.968002 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" event={"ID":"56453fbc-930c-4145-9104-ef8e74f89b85","Type":"ContainerStarted","Data":"27406f1ad862479106c9ed91bc8d3b599a0db4869eee3c3bca721b3b2b34e2fa"} Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.968045 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" event={"ID":"56453fbc-930c-4145-9104-ef8e74f89b85","Type":"ContainerStarted","Data":"77d94d436927d52a68cecd7715348e068926c6500913de946185410e37f4ebd0"} Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.968783 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:31 crc kubenswrapper[4669]: I1010 09:17:31.993940 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" podStartSLOduration=0.993920359 podStartE2EDuration="993.920359ms" podCreationTimestamp="2025-10-10 09:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:17:31.987354796 +0000 UTC m=+395.003373538" watchObservedRunningTime="2025-10-10 09:17:31.993920359 +0000 UTC m=+395.009939111" Oct 10 09:17:51 crc kubenswrapper[4669]: I1010 09:17:51.628101 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-c7s9q" Oct 10 09:17:51 crc kubenswrapper[4669]: I1010 09:17:51.678204 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w8ngm"] Oct 10 09:17:54 crc kubenswrapper[4669]: I1010 09:17:54.274540 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:17:54 crc kubenswrapper[4669]: I1010 09:17:54.274891 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:17:54 crc kubenswrapper[4669]: I1010 09:17:54.274935 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:17:54 crc kubenswrapper[4669]: I1010 09:17:54.275420 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"656cfe9a9a1a95c5e47506e7b135dc67d44e4e97a869b1e6e14c485ee25b4511"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:17:54 crc kubenswrapper[4669]: I1010 09:17:54.275461 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://656cfe9a9a1a95c5e47506e7b135dc67d44e4e97a869b1e6e14c485ee25b4511" gracePeriod=600 Oct 10 09:17:55 crc kubenswrapper[4669]: I1010 09:17:55.087563 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="656cfe9a9a1a95c5e47506e7b135dc67d44e4e97a869b1e6e14c485ee25b4511" exitCode=0 Oct 10 09:17:55 crc kubenswrapper[4669]: I1010 09:17:55.087802 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"656cfe9a9a1a95c5e47506e7b135dc67d44e4e97a869b1e6e14c485ee25b4511"} Oct 10 09:17:55 crc kubenswrapper[4669]: I1010 09:17:55.087981 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"86ae923b6b0c46f8077ccd2f9c57221fedfc20fa3778f2e007aa9fa0736e2d76"} Oct 10 09:17:55 crc kubenswrapper[4669]: I1010 09:17:55.088011 4669 scope.go:117] "RemoveContainer" containerID="2fd63ec848d52678ee7e05aa7a1c13461b3f93b657a2cfeb068d0b07d832c26c" Oct 10 09:18:16 crc kubenswrapper[4669]: I1010 09:18:16.713365 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" podUID="bfb7109a-2f1c-49c9-960b-0be9a7028b35" containerName="registry" containerID="cri-o://2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660" gracePeriod=30 Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.094234 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169031 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-tls\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169137 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfb7109a-2f1c-49c9-960b-0be9a7028b35-installation-pull-secrets\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169184 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-bound-sa-token\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169217 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-certificates\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169258 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfb7109a-2f1c-49c9-960b-0be9a7028b35-ca-trust-extracted\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169279 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-trusted-ca\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169393 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.169447 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbmhr\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-kube-api-access-zbmhr\") pod \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\" (UID: \"bfb7109a-2f1c-49c9-960b-0be9a7028b35\") " Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.170323 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.170450 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.175006 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.175952 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-kube-api-access-zbmhr" (OuterVolumeSpecName: "kube-api-access-zbmhr") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "kube-api-access-zbmhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.176018 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb7109a-2f1c-49c9-960b-0be9a7028b35-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.184578 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.188012 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfb7109a-2f1c-49c9-960b-0be9a7028b35-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.202857 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bfb7109a-2f1c-49c9-960b-0be9a7028b35" (UID: "bfb7109a-2f1c-49c9-960b-0be9a7028b35"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.212265 4669 generic.go:334] "Generic (PLEG): container finished" podID="bfb7109a-2f1c-49c9-960b-0be9a7028b35" containerID="2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660" exitCode=0 Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.212305 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" event={"ID":"bfb7109a-2f1c-49c9-960b-0be9a7028b35","Type":"ContainerDied","Data":"2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660"} Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.212316 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.212330 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w8ngm" event={"ID":"bfb7109a-2f1c-49c9-960b-0be9a7028b35","Type":"ContainerDied","Data":"11cb1eb9a2d49fcf5a76649bf6cb74d47fc4b06ae2819a2457d1019bc4b6eed2"} Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.212345 4669 scope.go:117] "RemoveContainer" containerID="2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.230946 4669 scope.go:117] "RemoveContainer" containerID="2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660" Oct 10 09:18:17 crc kubenswrapper[4669]: E1010 09:18:17.231339 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660\": container with ID starting with 2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660 not found: ID does not exist" containerID="2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.231377 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660"} err="failed to get container status \"2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660\": rpc error: code = NotFound desc = could not find container \"2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660\": container with ID starting with 2e424e918bac8aef08182648469313d1c8c52426145ccdc6b8f047bc950b5660 not found: ID does not exist" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.241660 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w8ngm"] Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.243202 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w8ngm"] Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.270933 4669 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bfb7109a-2f1c-49c9-960b-0be9a7028b35-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.270978 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.270991 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbmhr\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-kube-api-access-zbmhr\") on node \"crc\" DevicePath \"\"" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.271007 4669 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.271018 4669 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bfb7109a-2f1c-49c9-960b-0be9a7028b35-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.271029 4669 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bfb7109a-2f1c-49c9-960b-0be9a7028b35-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.271040 4669 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bfb7109a-2f1c-49c9-960b-0be9a7028b35-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 10 09:18:17 crc kubenswrapper[4669]: I1010 09:18:17.807742 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb7109a-2f1c-49c9-960b-0be9a7028b35" path="/var/lib/kubelet/pods/bfb7109a-2f1c-49c9-960b-0be9a7028b35/volumes" Oct 10 09:19:54 crc kubenswrapper[4669]: I1010 09:19:54.275003 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:19:54 crc kubenswrapper[4669]: I1010 09:19:54.275784 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:20:24 crc kubenswrapper[4669]: I1010 09:20:24.275401 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:20:24 crc kubenswrapper[4669]: I1010 09:20:24.276506 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:20:54 crc kubenswrapper[4669]: I1010 09:20:54.275200 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:20:54 crc kubenswrapper[4669]: I1010 09:20:54.275711 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:20:54 crc kubenswrapper[4669]: I1010 09:20:54.275765 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:20:54 crc kubenswrapper[4669]: I1010 09:20:54.276357 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"86ae923b6b0c46f8077ccd2f9c57221fedfc20fa3778f2e007aa9fa0736e2d76"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:20:54 crc kubenswrapper[4669]: I1010 09:20:54.276414 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://86ae923b6b0c46f8077ccd2f9c57221fedfc20fa3778f2e007aa9fa0736e2d76" gracePeriod=600 Oct 10 09:20:55 crc kubenswrapper[4669]: I1010 09:20:55.099746 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="86ae923b6b0c46f8077ccd2f9c57221fedfc20fa3778f2e007aa9fa0736e2d76" exitCode=0 Oct 10 09:20:55 crc kubenswrapper[4669]: I1010 09:20:55.099809 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"86ae923b6b0c46f8077ccd2f9c57221fedfc20fa3778f2e007aa9fa0736e2d76"} Oct 10 09:20:55 crc kubenswrapper[4669]: I1010 09:20:55.100335 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"b1be725249e1775acac8f350a1439ffc4d89a22bc933af6755f8ceefcd7eb018"} Oct 10 09:20:55 crc kubenswrapper[4669]: I1010 09:20:55.100357 4669 scope.go:117] "RemoveContainer" containerID="656cfe9a9a1a95c5e47506e7b135dc67d44e4e97a869b1e6e14c485ee25b4511" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.899762 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-45l5f"] Oct 10 09:21:23 crc kubenswrapper[4669]: E1010 09:21:23.901614 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb7109a-2f1c-49c9-960b-0be9a7028b35" containerName="registry" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.901712 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb7109a-2f1c-49c9-960b-0be9a7028b35" containerName="registry" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.901936 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb7109a-2f1c-49c9-960b-0be9a7028b35" containerName="registry" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.902456 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.905741 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-tqwjt"] Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.905838 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.906032 4669 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7m5p4" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.906154 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.906419 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-tqwjt" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.907936 4669 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-wlwzp" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.914986 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-45l5f"] Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.922685 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4258m"] Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.924071 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.929556 4669 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-fh2w6" Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.937567 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4258m"] Oct 10 09:21:23 crc kubenswrapper[4669]: I1010 09:21:23.943119 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-tqwjt"] Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.079218 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhkgr\" (UniqueName: \"kubernetes.io/projected/a9cc3175-a61f-44c4-8668-6270126b84b7-kube-api-access-xhkgr\") pod \"cert-manager-cainjector-7f985d654d-45l5f\" (UID: \"a9cc3175-a61f-44c4-8668-6270126b84b7\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.079355 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8ssp\" (UniqueName: \"kubernetes.io/projected/302f9809-912d-4e36-b0d5-7bdbae70071a-kube-api-access-j8ssp\") pod \"cert-manager-webhook-5655c58dd6-4258m\" (UID: \"302f9809-912d-4e36-b0d5-7bdbae70071a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.079395 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbwk7\" (UniqueName: \"kubernetes.io/projected/4ab4da17-72b9-45f6-94c3-75905972615a-kube-api-access-mbwk7\") pod \"cert-manager-5b446d88c5-tqwjt\" (UID: \"4ab4da17-72b9-45f6-94c3-75905972615a\") " pod="cert-manager/cert-manager-5b446d88c5-tqwjt" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.180980 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhkgr\" (UniqueName: \"kubernetes.io/projected/a9cc3175-a61f-44c4-8668-6270126b84b7-kube-api-access-xhkgr\") pod \"cert-manager-cainjector-7f985d654d-45l5f\" (UID: \"a9cc3175-a61f-44c4-8668-6270126b84b7\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.181293 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8ssp\" (UniqueName: \"kubernetes.io/projected/302f9809-912d-4e36-b0d5-7bdbae70071a-kube-api-access-j8ssp\") pod \"cert-manager-webhook-5655c58dd6-4258m\" (UID: \"302f9809-912d-4e36-b0d5-7bdbae70071a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.181484 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbwk7\" (UniqueName: \"kubernetes.io/projected/4ab4da17-72b9-45f6-94c3-75905972615a-kube-api-access-mbwk7\") pod \"cert-manager-5b446d88c5-tqwjt\" (UID: \"4ab4da17-72b9-45f6-94c3-75905972615a\") " pod="cert-manager/cert-manager-5b446d88c5-tqwjt" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.201765 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8ssp\" (UniqueName: \"kubernetes.io/projected/302f9809-912d-4e36-b0d5-7bdbae70071a-kube-api-access-j8ssp\") pod \"cert-manager-webhook-5655c58dd6-4258m\" (UID: \"302f9809-912d-4e36-b0d5-7bdbae70071a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.207826 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbwk7\" (UniqueName: \"kubernetes.io/projected/4ab4da17-72b9-45f6-94c3-75905972615a-kube-api-access-mbwk7\") pod \"cert-manager-5b446d88c5-tqwjt\" (UID: \"4ab4da17-72b9-45f6-94c3-75905972615a\") " pod="cert-manager/cert-manager-5b446d88c5-tqwjt" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.210540 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhkgr\" (UniqueName: \"kubernetes.io/projected/a9cc3175-a61f-44c4-8668-6270126b84b7-kube-api-access-xhkgr\") pod \"cert-manager-cainjector-7f985d654d-45l5f\" (UID: \"a9cc3175-a61f-44c4-8668-6270126b84b7\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.222728 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.230975 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-tqwjt" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.241019 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.645766 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-45l5f"] Oct 10 09:21:24 crc kubenswrapper[4669]: W1010 09:21:24.655167 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9cc3175_a61f_44c4_8668_6270126b84b7.slice/crio-386b98d2712a689b11c1145aa2e01a2e3333b184fdc57c24db96a62db6d5205d WatchSource:0}: Error finding container 386b98d2712a689b11c1145aa2e01a2e3333b184fdc57c24db96a62db6d5205d: Status 404 returned error can't find the container with id 386b98d2712a689b11c1145aa2e01a2e3333b184fdc57c24db96a62db6d5205d Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.660136 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.699011 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-tqwjt"] Oct 10 09:21:24 crc kubenswrapper[4669]: W1010 09:21:24.703792 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ab4da17_72b9_45f6_94c3_75905972615a.slice/crio-894cad6aa7b697c0600303a676beb9ce0ffb58537e8aeae90a6f56eb1879d265 WatchSource:0}: Error finding container 894cad6aa7b697c0600303a676beb9ce0ffb58537e8aeae90a6f56eb1879d265: Status 404 returned error can't find the container with id 894cad6aa7b697c0600303a676beb9ce0ffb58537e8aeae90a6f56eb1879d265 Oct 10 09:21:24 crc kubenswrapper[4669]: I1010 09:21:24.731737 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4258m"] Oct 10 09:21:24 crc kubenswrapper[4669]: W1010 09:21:24.733830 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302f9809_912d_4e36_b0d5_7bdbae70071a.slice/crio-32f9321fbf3349c4f5d567cd0da4110e09be6ef6ddacc6583e8866380fd05645 WatchSource:0}: Error finding container 32f9321fbf3349c4f5d567cd0da4110e09be6ef6ddacc6583e8866380fd05645: Status 404 returned error can't find the container with id 32f9321fbf3349c4f5d567cd0da4110e09be6ef6ddacc6583e8866380fd05645 Oct 10 09:21:25 crc kubenswrapper[4669]: I1010 09:21:25.259825 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-tqwjt" event={"ID":"4ab4da17-72b9-45f6-94c3-75905972615a","Type":"ContainerStarted","Data":"894cad6aa7b697c0600303a676beb9ce0ffb58537e8aeae90a6f56eb1879d265"} Oct 10 09:21:25 crc kubenswrapper[4669]: I1010 09:21:25.265680 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" event={"ID":"302f9809-912d-4e36-b0d5-7bdbae70071a","Type":"ContainerStarted","Data":"32f9321fbf3349c4f5d567cd0da4110e09be6ef6ddacc6583e8866380fd05645"} Oct 10 09:21:25 crc kubenswrapper[4669]: I1010 09:21:25.267077 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" event={"ID":"a9cc3175-a61f-44c4-8668-6270126b84b7","Type":"ContainerStarted","Data":"386b98d2712a689b11c1145aa2e01a2e3333b184fdc57c24db96a62db6d5205d"} Oct 10 09:21:29 crc kubenswrapper[4669]: I1010 09:21:29.294009 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-tqwjt" event={"ID":"4ab4da17-72b9-45f6-94c3-75905972615a","Type":"ContainerStarted","Data":"dfc630952443f0aa99a2b7394112fb0a7c610893864023392c8ae0b4e4862a12"} Oct 10 09:21:29 crc kubenswrapper[4669]: I1010 09:21:29.295172 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" event={"ID":"302f9809-912d-4e36-b0d5-7bdbae70071a","Type":"ContainerStarted","Data":"5981e31fa6335b93c2ffe50eca90f809586fd223974dd791a0cb7f88fbed384c"} Oct 10 09:21:29 crc kubenswrapper[4669]: I1010 09:21:29.295249 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" Oct 10 09:21:29 crc kubenswrapper[4669]: I1010 09:21:29.296771 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" event={"ID":"a9cc3175-a61f-44c4-8668-6270126b84b7","Type":"ContainerStarted","Data":"5d7ae496d88e0960113a186f7084075ee37d03a1ec1a416d08de00b9a1ca566a"} Oct 10 09:21:29 crc kubenswrapper[4669]: I1010 09:21:29.309986 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-tqwjt" podStartSLOduration=2.739469313 podStartE2EDuration="6.309966685s" podCreationTimestamp="2025-10-10 09:21:23 +0000 UTC" firstStartedPulling="2025-10-10 09:21:24.705976942 +0000 UTC m=+627.721995684" lastFinishedPulling="2025-10-10 09:21:28.276474304 +0000 UTC m=+631.292493056" observedRunningTime="2025-10-10 09:21:29.306230453 +0000 UTC m=+632.322249195" watchObservedRunningTime="2025-10-10 09:21:29.309966685 +0000 UTC m=+632.325985437" Oct 10 09:21:29 crc kubenswrapper[4669]: I1010 09:21:29.337276 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" podStartSLOduration=2.781186404 podStartE2EDuration="6.337252749s" podCreationTimestamp="2025-10-10 09:21:23 +0000 UTC" firstStartedPulling="2025-10-10 09:21:24.735800668 +0000 UTC m=+627.751819410" lastFinishedPulling="2025-10-10 09:21:28.291867003 +0000 UTC m=+631.307885755" observedRunningTime="2025-10-10 09:21:29.333738995 +0000 UTC m=+632.349757757" watchObservedRunningTime="2025-10-10 09:21:29.337252749 +0000 UTC m=+632.353271501" Oct 10 09:21:29 crc kubenswrapper[4669]: I1010 09:21:29.348716 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-45l5f" podStartSLOduration=2.72550287 podStartE2EDuration="6.34869867s" podCreationTimestamp="2025-10-10 09:21:23 +0000 UTC" firstStartedPulling="2025-10-10 09:21:24.659847656 +0000 UTC m=+627.675866388" lastFinishedPulling="2025-10-10 09:21:28.283043436 +0000 UTC m=+631.299062188" observedRunningTime="2025-10-10 09:21:29.3483873 +0000 UTC m=+632.364406042" watchObservedRunningTime="2025-10-10 09:21:29.34869867 +0000 UTC m=+632.364717412" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.243163 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-4258m" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.527740 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gbsxj"] Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.528087 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-controller" containerID="cri-o://7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.528194 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-acl-logging" containerID="cri-o://1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.528202 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="nbdb" containerID="cri-o://8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.528246 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="sbdb" containerID="cri-o://7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.528265 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="northd" containerID="cri-o://167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.528290 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.528253 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-node" containerID="cri-o://7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.571119 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" containerID="cri-o://a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" gracePeriod=30 Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.878742 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/3.log" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.881336 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovn-acl-logging/0.log" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.881889 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovn-controller/0.log" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.882236 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955506 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-48gj4"] Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955729 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="northd" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955742 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="northd" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955755 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955761 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955766 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-acl-logging" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955773 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-acl-logging" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955780 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="sbdb" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955785 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="sbdb" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955793 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="nbdb" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955798 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="nbdb" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955808 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955813 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955819 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955826 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955834 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955841 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955850 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kubecfg-setup" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955856 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kubecfg-setup" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955865 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-node" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955871 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-node" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.955879 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955885 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955973 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955985 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.955994 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="kube-rbac-proxy-node" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956006 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-acl-logging" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956016 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="northd" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956024 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovn-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956032 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956038 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="nbdb" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956046 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="sbdb" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956053 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956061 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.956150 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956158 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: E1010 09:21:34.956174 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956183 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.956281 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerName="ovnkube-controller" Oct 10 09:21:34 crc kubenswrapper[4669]: I1010 09:21:34.957803 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.030555 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-netns\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.030933 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-ovn-kubernetes\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031139 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-var-lib-openvswitch\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031225 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-systemd-units\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.030749 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031000 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031278 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031408 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031319 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovn-node-metrics-cert\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031568 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-netd\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031671 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-systemd\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031704 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031732 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-bin\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031766 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xsrl\" (UniqueName: \"kubernetes.io/projected/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-kube-api-access-6xsrl\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031799 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-slash\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031827 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-script-lib\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031863 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-env-overrides\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031882 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-etc-openvswitch\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031944 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-log-socket\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031967 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-openvswitch\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.031983 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-ovn\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032012 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-config\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032032 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-kubelet\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032065 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-node-log\") pod \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\" (UID: \"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e\") " Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032661 4669 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032694 4669 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032705 4669 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032717 4669 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032756 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-node-log" (OuterVolumeSpecName: "node-log") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032878 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.032918 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033120 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033165 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033190 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-log-socket" (OuterVolumeSpecName: "log-socket") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033209 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033228 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033321 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033352 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033377 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033536 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-slash" (OuterVolumeSpecName: "host-slash") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.033606 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.039662 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.040154 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-kube-api-access-6xsrl" (OuterVolumeSpecName: "kube-api-access-6xsrl") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "kube-api-access-6xsrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.049734 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" (UID: "e1b02e9a-7e2e-473d-a810-d4ece0d3a18e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133377 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-kubelet\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-ovnkube-config\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133519 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133620 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-systemd\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133679 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-env-overrides\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133702 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-cni-netd\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133722 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133760 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-slash\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133780 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-run-ovn-kubernetes\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133804 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5825b115-a79f-4430-aa58-48494458cedd-ovn-node-metrics-cert\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133864 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-ovn\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133928 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-node-log\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.133963 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-systemd-units\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134005 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-var-lib-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134044 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-etc-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134064 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-log-socket\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134104 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-cni-bin\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134175 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk5mj\" (UniqueName: \"kubernetes.io/projected/5825b115-a79f-4430-aa58-48494458cedd-kube-api-access-zk5mj\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134218 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-run-netns\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134263 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-ovnkube-script-lib\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134337 4669 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134360 4669 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-node-log\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134371 4669 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134381 4669 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134390 4669 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134400 4669 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134420 4669 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134429 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xsrl\" (UniqueName: \"kubernetes.io/projected/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-kube-api-access-6xsrl\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134438 4669 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-host-slash\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134449 4669 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134459 4669 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134469 4669 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134519 4669 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-log-socket\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134538 4669 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134551 4669 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.134646 4669 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236282 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk5mj\" (UniqueName: \"kubernetes.io/projected/5825b115-a79f-4430-aa58-48494458cedd-kube-api-access-zk5mj\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236337 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-run-netns\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236374 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-ovnkube-script-lib\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236410 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-kubelet\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236430 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-ovnkube-config\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236470 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236495 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-systemd\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236515 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-env-overrides\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236534 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-cni-netd\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236550 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236571 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-slash\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236649 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-run-ovn-kubernetes\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236663 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5825b115-a79f-4430-aa58-48494458cedd-ovn-node-metrics-cert\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236678 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-ovn\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236693 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-node-log\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236717 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-systemd-units\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236733 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-var-lib-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236747 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-etc-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236768 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-log-socket\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236784 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-cni-bin\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236786 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-systemd\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236866 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-kubelet\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236879 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-cni-bin\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.236951 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-run-netns\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237035 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237072 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-ovn\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237102 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-cni-netd\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237129 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-run-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237156 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-slash\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237188 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-host-run-ovn-kubernetes\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237355 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-var-lib-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237426 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-node-log\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237490 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-systemd-units\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237525 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-etc-openvswitch\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237597 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5825b115-a79f-4430-aa58-48494458cedd-log-socket\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237752 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-env-overrides\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237820 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-ovnkube-script-lib\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.237903 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5825b115-a79f-4430-aa58-48494458cedd-ovnkube-config\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.242030 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5825b115-a79f-4430-aa58-48494458cedd-ovn-node-metrics-cert\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.252436 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk5mj\" (UniqueName: \"kubernetes.io/projected/5825b115-a79f-4430-aa58-48494458cedd-kube-api-access-zk5mj\") pod \"ovnkube-node-48gj4\" (UID: \"5825b115-a79f-4430-aa58-48494458cedd\") " pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.272260 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.334924 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/2.log" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.335564 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/1.log" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.335636 4669 generic.go:334] "Generic (PLEG): container finished" podID="9502d677-e546-4df7-96c2-bce8274c0f57" containerID="7b2f37ba6882cf12f83ece0045d1343775e215b0d1ebf073935817d0e74dc426" exitCode=2 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.335734 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerDied","Data":"7b2f37ba6882cf12f83ece0045d1343775e215b0d1ebf073935817d0e74dc426"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.335811 4669 scope.go:117] "RemoveContainer" containerID="7f161aa411a94813f94ffde4694cf9d2656855320d770e8c41e9489fd08c8ad7" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.336413 4669 scope.go:117] "RemoveContainer" containerID="7b2f37ba6882cf12f83ece0045d1343775e215b0d1ebf073935817d0e74dc426" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.336659 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nq6jn_openshift-multus(9502d677-e546-4df7-96c2-bce8274c0f57)\"" pod="openshift-multus/multus-nq6jn" podUID="9502d677-e546-4df7-96c2-bce8274c0f57" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.336921 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"a11afa127f41f327c710ee791d653ce848f51c36b0d17c5d46eb151480d08409"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.340526 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovnkube-controller/3.log" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.343426 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovn-acl-logging/0.log" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.343969 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gbsxj_e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/ovn-controller/0.log" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344479 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" exitCode=0 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344517 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" exitCode=0 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344526 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" exitCode=0 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344535 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" exitCode=0 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344543 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" exitCode=0 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344551 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" exitCode=0 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344559 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" exitCode=143 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344568 4669 generic.go:334] "Generic (PLEG): container finished" podID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" containerID="7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" exitCode=143 Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344615 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344644 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344657 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344667 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344685 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344703 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344713 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344724 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344729 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344735 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344740 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344747 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344752 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344758 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344762 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344767 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344776 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344783 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344789 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344795 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344800 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344806 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344811 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344816 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344821 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344827 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344833 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344840 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344847 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344853 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344858 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344863 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344869 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344873 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344891 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344896 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344901 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344906 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344913 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" event={"ID":"e1b02e9a-7e2e-473d-a810-d4ece0d3a18e","Type":"ContainerDied","Data":"732d505dfba05c3478bc1aba7ec5bed96b5acf93eabbdceef4e9c76b0307bba5"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344921 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344926 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344931 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344937 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344941 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344946 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344951 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344956 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344960 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.344966 4669 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.345082 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gbsxj" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.376869 4669 scope.go:117] "RemoveContainer" containerID="a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.390325 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gbsxj"] Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.395837 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gbsxj"] Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.398190 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.411943 4669 scope.go:117] "RemoveContainer" containerID="7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.427871 4669 scope.go:117] "RemoveContainer" containerID="8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.472842 4669 scope.go:117] "RemoveContainer" containerID="167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.485522 4669 scope.go:117] "RemoveContainer" containerID="455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.497658 4669 scope.go:117] "RemoveContainer" containerID="7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.509292 4669 scope.go:117] "RemoveContainer" containerID="1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.527786 4669 scope.go:117] "RemoveContainer" containerID="7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.551925 4669 scope.go:117] "RemoveContainer" containerID="a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.570908 4669 scope.go:117] "RemoveContainer" containerID="a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.571322 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": container with ID starting with a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d not found: ID does not exist" containerID="a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.571369 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} err="failed to get container status \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": rpc error: code = NotFound desc = could not find container \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": container with ID starting with a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.571404 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.573221 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": container with ID starting with fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b not found: ID does not exist" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.573262 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} err="failed to get container status \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": rpc error: code = NotFound desc = could not find container \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": container with ID starting with fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.573288 4669 scope.go:117] "RemoveContainer" containerID="7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.573624 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": container with ID starting with 7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364 not found: ID does not exist" containerID="7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.573649 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} err="failed to get container status \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": rpc error: code = NotFound desc = could not find container \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": container with ID starting with 7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.573662 4669 scope.go:117] "RemoveContainer" containerID="8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.574061 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": container with ID starting with 8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec not found: ID does not exist" containerID="8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.574086 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} err="failed to get container status \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": rpc error: code = NotFound desc = could not find container \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": container with ID starting with 8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.574105 4669 scope.go:117] "RemoveContainer" containerID="167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.574426 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": container with ID starting with 167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1 not found: ID does not exist" containerID="167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.574452 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} err="failed to get container status \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": rpc error: code = NotFound desc = could not find container \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": container with ID starting with 167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.574465 4669 scope.go:117] "RemoveContainer" containerID="455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.574996 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": container with ID starting with 455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b not found: ID does not exist" containerID="455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.575023 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} err="failed to get container status \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": rpc error: code = NotFound desc = could not find container \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": container with ID starting with 455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.575041 4669 scope.go:117] "RemoveContainer" containerID="7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.575451 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": container with ID starting with 7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a not found: ID does not exist" containerID="7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.575474 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} err="failed to get container status \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": rpc error: code = NotFound desc = could not find container \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": container with ID starting with 7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.575490 4669 scope.go:117] "RemoveContainer" containerID="1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.575742 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": container with ID starting with 1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653 not found: ID does not exist" containerID="1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.575768 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} err="failed to get container status \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": rpc error: code = NotFound desc = could not find container \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": container with ID starting with 1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.575789 4669 scope.go:117] "RemoveContainer" containerID="7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.576063 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": container with ID starting with 7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29 not found: ID does not exist" containerID="7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.576098 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} err="failed to get container status \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": rpc error: code = NotFound desc = could not find container \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": container with ID starting with 7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.576120 4669 scope.go:117] "RemoveContainer" containerID="a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c" Oct 10 09:21:35 crc kubenswrapper[4669]: E1010 09:21:35.576365 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": container with ID starting with a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c not found: ID does not exist" containerID="a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.576389 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} err="failed to get container status \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": rpc error: code = NotFound desc = could not find container \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": container with ID starting with a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.576404 4669 scope.go:117] "RemoveContainer" containerID="a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.576856 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} err="failed to get container status \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": rpc error: code = NotFound desc = could not find container \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": container with ID starting with a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.576880 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.577131 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} err="failed to get container status \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": rpc error: code = NotFound desc = could not find container \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": container with ID starting with fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.577152 4669 scope.go:117] "RemoveContainer" containerID="7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.577500 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} err="failed to get container status \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": rpc error: code = NotFound desc = could not find container \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": container with ID starting with 7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.577525 4669 scope.go:117] "RemoveContainer" containerID="8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.577929 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} err="failed to get container status \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": rpc error: code = NotFound desc = could not find container \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": container with ID starting with 8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.577949 4669 scope.go:117] "RemoveContainer" containerID="167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.578202 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} err="failed to get container status \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": rpc error: code = NotFound desc = could not find container \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": container with ID starting with 167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.578230 4669 scope.go:117] "RemoveContainer" containerID="455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.578448 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} err="failed to get container status \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": rpc error: code = NotFound desc = could not find container \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": container with ID starting with 455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.578469 4669 scope.go:117] "RemoveContainer" containerID="7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.578841 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} err="failed to get container status \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": rpc error: code = NotFound desc = could not find container \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": container with ID starting with 7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.578862 4669 scope.go:117] "RemoveContainer" containerID="1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579064 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} err="failed to get container status \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": rpc error: code = NotFound desc = could not find container \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": container with ID starting with 1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579090 4669 scope.go:117] "RemoveContainer" containerID="7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579361 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} err="failed to get container status \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": rpc error: code = NotFound desc = could not find container \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": container with ID starting with 7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579385 4669 scope.go:117] "RemoveContainer" containerID="a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579663 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} err="failed to get container status \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": rpc error: code = NotFound desc = could not find container \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": container with ID starting with a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579690 4669 scope.go:117] "RemoveContainer" containerID="a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579907 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} err="failed to get container status \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": rpc error: code = NotFound desc = could not find container \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": container with ID starting with a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.579935 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580184 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} err="failed to get container status \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": rpc error: code = NotFound desc = could not find container \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": container with ID starting with fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580211 4669 scope.go:117] "RemoveContainer" containerID="7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580411 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} err="failed to get container status \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": rpc error: code = NotFound desc = could not find container \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": container with ID starting with 7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580433 4669 scope.go:117] "RemoveContainer" containerID="8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580630 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} err="failed to get container status \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": rpc error: code = NotFound desc = could not find container \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": container with ID starting with 8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580654 4669 scope.go:117] "RemoveContainer" containerID="167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580809 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} err="failed to get container status \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": rpc error: code = NotFound desc = could not find container \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": container with ID starting with 167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.580827 4669 scope.go:117] "RemoveContainer" containerID="455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581003 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} err="failed to get container status \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": rpc error: code = NotFound desc = could not find container \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": container with ID starting with 455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581027 4669 scope.go:117] "RemoveContainer" containerID="7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581217 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} err="failed to get container status \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": rpc error: code = NotFound desc = could not find container \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": container with ID starting with 7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581234 4669 scope.go:117] "RemoveContainer" containerID="1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581618 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} err="failed to get container status \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": rpc error: code = NotFound desc = could not find container \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": container with ID starting with 1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581643 4669 scope.go:117] "RemoveContainer" containerID="7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581898 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} err="failed to get container status \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": rpc error: code = NotFound desc = could not find container \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": container with ID starting with 7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.581922 4669 scope.go:117] "RemoveContainer" containerID="a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.582106 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} err="failed to get container status \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": rpc error: code = NotFound desc = could not find container \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": container with ID starting with a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.582125 4669 scope.go:117] "RemoveContainer" containerID="a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.582468 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d"} err="failed to get container status \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": rpc error: code = NotFound desc = could not find container \"a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d\": container with ID starting with a8db27c595b9b2962aa541b5a342cc524d0383565a95aeae29ae7e364544203d not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.582492 4669 scope.go:117] "RemoveContainer" containerID="fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.582769 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b"} err="failed to get container status \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": rpc error: code = NotFound desc = could not find container \"fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b\": container with ID starting with fb58fc01af32a6c8a48be63077b64b9a015cfa5c6909c843560efbf563dcab4b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.582791 4669 scope.go:117] "RemoveContainer" containerID="7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.583015 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364"} err="failed to get container status \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": rpc error: code = NotFound desc = could not find container \"7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364\": container with ID starting with 7e0a385d04860a88f050425ed5ea1f7dd2471db67c63a841f8e40760de8d7364 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.583036 4669 scope.go:117] "RemoveContainer" containerID="8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.583491 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec"} err="failed to get container status \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": rpc error: code = NotFound desc = could not find container \"8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec\": container with ID starting with 8e186fa977b8220171c63c577bd13cb7430dfcb74b3e5e276ca4fa140f3998ec not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.583516 4669 scope.go:117] "RemoveContainer" containerID="167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.583840 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1"} err="failed to get container status \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": rpc error: code = NotFound desc = could not find container \"167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1\": container with ID starting with 167f828fdc4f2dbe6175c66e03ab7e60c01c4a11639396275f1460410bb150c1 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.583861 4669 scope.go:117] "RemoveContainer" containerID="455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.584249 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b"} err="failed to get container status \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": rpc error: code = NotFound desc = could not find container \"455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b\": container with ID starting with 455e7e1db7c4f43119551253f92a2d3e8c77e8a5e37b652f476e15641e6c668b not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.584271 4669 scope.go:117] "RemoveContainer" containerID="7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.584602 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a"} err="failed to get container status \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": rpc error: code = NotFound desc = could not find container \"7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a\": container with ID starting with 7b8be786f5fcac4c6332c32a4e15688611c04f2a26a34f3ca0491909e1c2867a not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.584639 4669 scope.go:117] "RemoveContainer" containerID="1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.584915 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653"} err="failed to get container status \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": rpc error: code = NotFound desc = could not find container \"1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653\": container with ID starting with 1fbfeda73d96623c43fd8ad7195d3c5cd998c482fa6ffc04df770b8735818653 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.584963 4669 scope.go:117] "RemoveContainer" containerID="7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.585290 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29"} err="failed to get container status \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": rpc error: code = NotFound desc = could not find container \"7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29\": container with ID starting with 7928de274bfc228637bb7ff4eebf5dab2c1b5c4f507b182d37c34d9f02fcfe29 not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.585314 4669 scope.go:117] "RemoveContainer" containerID="a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.585717 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c"} err="failed to get container status \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": rpc error: code = NotFound desc = could not find container \"a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c\": container with ID starting with a1082ea761b00621760d3305c10b40f5dc56c5570e274c57b63461e6921f2b3c not found: ID does not exist" Oct 10 09:21:35 crc kubenswrapper[4669]: I1010 09:21:35.802118 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1b02e9a-7e2e-473d-a810-d4ece0d3a18e" path="/var/lib/kubelet/pods/e1b02e9a-7e2e-473d-a810-d4ece0d3a18e/volumes" Oct 10 09:21:36 crc kubenswrapper[4669]: I1010 09:21:36.352283 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/2.log" Oct 10 09:21:36 crc kubenswrapper[4669]: I1010 09:21:36.353473 4669 generic.go:334] "Generic (PLEG): container finished" podID="5825b115-a79f-4430-aa58-48494458cedd" containerID="0eff753f70685d00a1dabe30d110d8d04ceb87db418ceb76f77b9f8ec3b2943b" exitCode=0 Oct 10 09:21:36 crc kubenswrapper[4669]: I1010 09:21:36.353507 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerDied","Data":"0eff753f70685d00a1dabe30d110d8d04ceb87db418ceb76f77b9f8ec3b2943b"} Oct 10 09:21:37 crc kubenswrapper[4669]: I1010 09:21:37.365405 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"c4a03d6dc093995e9e6490b3c07db942f3bc2449b908fc3afd74fcf3a1b297f9"} Oct 10 09:21:37 crc kubenswrapper[4669]: I1010 09:21:37.366294 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"6baebde75d563b1aa90cfddf0719053274b120a4c0d7ad86b8072002eed16fc2"} Oct 10 09:21:37 crc kubenswrapper[4669]: I1010 09:21:37.366360 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"bd5cf0fafb52802bda081b3af63235b9152c790516a6bbcd84454553a45b196b"} Oct 10 09:21:37 crc kubenswrapper[4669]: I1010 09:21:37.366419 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"9e5a9cde248f7a6d858ba32207a49298ff9a5b4029fc48786f070f879fdaf6a6"} Oct 10 09:21:37 crc kubenswrapper[4669]: I1010 09:21:37.366478 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"b0c423cad944a104745dd2f4202c2eed90a01173ddd34b488e4cad698b9c3472"} Oct 10 09:21:37 crc kubenswrapper[4669]: I1010 09:21:37.366530 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"dbb0981c80bd9ddc32a0d769afd40a913af7e11798521e38474e75915156793d"} Oct 10 09:21:39 crc kubenswrapper[4669]: I1010 09:21:39.380511 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"f5ca18022ffccff8e97cb3a43f04e969f4bfc2066d41e22d1cae9053b2e14d58"} Oct 10 09:21:41 crc kubenswrapper[4669]: I1010 09:21:41.394857 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" event={"ID":"5825b115-a79f-4430-aa58-48494458cedd","Type":"ContainerStarted","Data":"f2512f67b86deea85535a0f1220ab256ca0ff83e05742cdce5271b74e290e9c1"} Oct 10 09:21:41 crc kubenswrapper[4669]: I1010 09:21:41.396128 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:41 crc kubenswrapper[4669]: I1010 09:21:41.396161 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:41 crc kubenswrapper[4669]: I1010 09:21:41.396214 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:41 crc kubenswrapper[4669]: I1010 09:21:41.423640 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" podStartSLOduration=7.423619955 podStartE2EDuration="7.423619955s" podCreationTimestamp="2025-10-10 09:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:21:41.422323922 +0000 UTC m=+644.438342674" watchObservedRunningTime="2025-10-10 09:21:41.423619955 +0000 UTC m=+644.439638727" Oct 10 09:21:41 crc kubenswrapper[4669]: I1010 09:21:41.429258 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:41 crc kubenswrapper[4669]: I1010 09:21:41.442813 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:21:49 crc kubenswrapper[4669]: I1010 09:21:49.795739 4669 scope.go:117] "RemoveContainer" containerID="7b2f37ba6882cf12f83ece0045d1343775e215b0d1ebf073935817d0e74dc426" Oct 10 09:21:49 crc kubenswrapper[4669]: E1010 09:21:49.796845 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nq6jn_openshift-multus(9502d677-e546-4df7-96c2-bce8274c0f57)\"" pod="openshift-multus/multus-nq6jn" podUID="9502d677-e546-4df7-96c2-bce8274c0f57" Oct 10 09:22:00 crc kubenswrapper[4669]: I1010 09:22:00.795457 4669 scope.go:117] "RemoveContainer" containerID="7b2f37ba6882cf12f83ece0045d1343775e215b0d1ebf073935817d0e74dc426" Oct 10 09:22:01 crc kubenswrapper[4669]: I1010 09:22:01.508411 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nq6jn_9502d677-e546-4df7-96c2-bce8274c0f57/kube-multus/2.log" Oct 10 09:22:01 crc kubenswrapper[4669]: I1010 09:22:01.508725 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nq6jn" event={"ID":"9502d677-e546-4df7-96c2-bce8274c0f57","Type":"ContainerStarted","Data":"6198270f1c4ae63d4fdcd9fb6c05c8dc73c3bb6d2a7060f634ca07242a02063f"} Oct 10 09:22:05 crc kubenswrapper[4669]: I1010 09:22:05.302228 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-48gj4" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.353428 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp"] Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.354959 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.357263 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.365670 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp"] Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.468275 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.468320 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.468346 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcmpx\" (UniqueName: \"kubernetes.io/projected/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-kube-api-access-mcmpx\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.569424 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.569481 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.569519 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcmpx\" (UniqueName: \"kubernetes.io/projected/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-kube-api-access-mcmpx\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.570408 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.570573 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.597216 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcmpx\" (UniqueName: \"kubernetes.io/projected/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-kube-api-access-mcmpx\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.672115 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:13 crc kubenswrapper[4669]: I1010 09:22:13.852603 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp"] Oct 10 09:22:13 crc kubenswrapper[4669]: W1010 09:22:13.859155 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15e5bfc4_8089_4e7f_9c96_66e5090b8dcc.slice/crio-9a917678925b42a519c4bff3d586438d530201ab5809aebc1c9846b8cc6e68bf WatchSource:0}: Error finding container 9a917678925b42a519c4bff3d586438d530201ab5809aebc1c9846b8cc6e68bf: Status 404 returned error can't find the container with id 9a917678925b42a519c4bff3d586438d530201ab5809aebc1c9846b8cc6e68bf Oct 10 09:22:14 crc kubenswrapper[4669]: I1010 09:22:14.583820 4669 generic.go:334] "Generic (PLEG): container finished" podID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerID="d79bd06ae2c1595095c5e76dd78842b8719ca1a48fc2318c1265a2d1ec8598ba" exitCode=0 Oct 10 09:22:14 crc kubenswrapper[4669]: I1010 09:22:14.583870 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" event={"ID":"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc","Type":"ContainerDied","Data":"d79bd06ae2c1595095c5e76dd78842b8719ca1a48fc2318c1265a2d1ec8598ba"} Oct 10 09:22:14 crc kubenswrapper[4669]: I1010 09:22:14.584124 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" event={"ID":"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc","Type":"ContainerStarted","Data":"9a917678925b42a519c4bff3d586438d530201ab5809aebc1c9846b8cc6e68bf"} Oct 10 09:22:16 crc kubenswrapper[4669]: I1010 09:22:16.598718 4669 generic.go:334] "Generic (PLEG): container finished" podID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerID="702603fcfb3532d00fd2da31dcf39aaf7d5871e383ebf930ef3403bc12b62d76" exitCode=0 Oct 10 09:22:16 crc kubenswrapper[4669]: I1010 09:22:16.598802 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" event={"ID":"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc","Type":"ContainerDied","Data":"702603fcfb3532d00fd2da31dcf39aaf7d5871e383ebf930ef3403bc12b62d76"} Oct 10 09:22:17 crc kubenswrapper[4669]: I1010 09:22:17.609622 4669 generic.go:334] "Generic (PLEG): container finished" podID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerID="a12bc9175efcafe40953138311b71f50a14f68d30f883624575fab6848b505a4" exitCode=0 Oct 10 09:22:17 crc kubenswrapper[4669]: I1010 09:22:17.609693 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" event={"ID":"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc","Type":"ContainerDied","Data":"a12bc9175efcafe40953138311b71f50a14f68d30f883624575fab6848b505a4"} Oct 10 09:22:18 crc kubenswrapper[4669]: I1010 09:22:18.937029 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.034680 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-bundle\") pod \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.034832 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcmpx\" (UniqueName: \"kubernetes.io/projected/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-kube-api-access-mcmpx\") pod \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.034916 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-util\") pod \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\" (UID: \"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc\") " Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.035356 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-bundle" (OuterVolumeSpecName: "bundle") pod "15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" (UID: "15e5bfc4-8089-4e7f-9c96-66e5090b8dcc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.039491 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-kube-api-access-mcmpx" (OuterVolumeSpecName: "kube-api-access-mcmpx") pod "15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" (UID: "15e5bfc4-8089-4e7f-9c96-66e5090b8dcc"). InnerVolumeSpecName "kube-api-access-mcmpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.065128 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-util" (OuterVolumeSpecName: "util") pod "15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" (UID: "15e5bfc4-8089-4e7f-9c96-66e5090b8dcc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.136656 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcmpx\" (UniqueName: \"kubernetes.io/projected/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-kube-api-access-mcmpx\") on node \"crc\" DevicePath \"\"" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.136702 4669 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-util\") on node \"crc\" DevicePath \"\"" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.136714 4669 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15e5bfc4-8089-4e7f-9c96-66e5090b8dcc-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.626172 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" event={"ID":"15e5bfc4-8089-4e7f-9c96-66e5090b8dcc","Type":"ContainerDied","Data":"9a917678925b42a519c4bff3d586438d530201ab5809aebc1c9846b8cc6e68bf"} Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.626441 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a917678925b42a519c4bff3d586438d530201ab5809aebc1c9846b8cc6e68bf" Oct 10 09:22:19 crc kubenswrapper[4669]: I1010 09:22:19.626253 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.166544 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv"] Oct 10 09:22:21 crc kubenswrapper[4669]: E1010 09:22:21.166811 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerName="extract" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.166828 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerName="extract" Oct 10 09:22:21 crc kubenswrapper[4669]: E1010 09:22:21.166845 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerName="pull" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.166852 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerName="pull" Oct 10 09:22:21 crc kubenswrapper[4669]: E1010 09:22:21.166868 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerName="util" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.166878 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerName="util" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.167000 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="15e5bfc4-8089-4e7f-9c96-66e5090b8dcc" containerName="extract" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.167448 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.168978 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8bwgb" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.169520 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.170233 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.179381 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv"] Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.363846 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzht8\" (UniqueName: \"kubernetes.io/projected/1b2674b7-3907-4f5a-b3b8-cc6945e4f004-kube-api-access-tzht8\") pod \"nmstate-operator-858ddd8f98-qwxrv\" (UID: \"1b2674b7-3907-4f5a-b3b8-cc6945e4f004\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.465440 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzht8\" (UniqueName: \"kubernetes.io/projected/1b2674b7-3907-4f5a-b3b8-cc6945e4f004-kube-api-access-tzht8\") pod \"nmstate-operator-858ddd8f98-qwxrv\" (UID: \"1b2674b7-3907-4f5a-b3b8-cc6945e4f004\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.486514 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzht8\" (UniqueName: \"kubernetes.io/projected/1b2674b7-3907-4f5a-b3b8-cc6945e4f004-kube-api-access-tzht8\") pod \"nmstate-operator-858ddd8f98-qwxrv\" (UID: \"1b2674b7-3907-4f5a-b3b8-cc6945e4f004\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" Oct 10 09:22:21 crc kubenswrapper[4669]: I1010 09:22:21.784968 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" Oct 10 09:22:22 crc kubenswrapper[4669]: I1010 09:22:22.000240 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv"] Oct 10 09:22:22 crc kubenswrapper[4669]: I1010 09:22:22.641525 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" event={"ID":"1b2674b7-3907-4f5a-b3b8-cc6945e4f004","Type":"ContainerStarted","Data":"38f3f970e41b179ec2875f71a72dcc451800f5e3aea5133dfb965301a2a742d7"} Oct 10 09:22:24 crc kubenswrapper[4669]: I1010 09:22:24.652928 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" event={"ID":"1b2674b7-3907-4f5a-b3b8-cc6945e4f004","Type":"ContainerStarted","Data":"8608b04d28c37d748f43fdb1f8384287d33ff46a25f23773112504fdb1de675e"} Oct 10 09:22:24 crc kubenswrapper[4669]: I1010 09:22:24.676292 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qwxrv" podStartSLOduration=1.321450119 podStartE2EDuration="3.676276104s" podCreationTimestamp="2025-10-10 09:22:21 +0000 UTC" firstStartedPulling="2025-10-10 09:22:22.063728523 +0000 UTC m=+685.079747265" lastFinishedPulling="2025-10-10 09:22:24.418554508 +0000 UTC m=+687.434573250" observedRunningTime="2025-10-10 09:22:24.671959444 +0000 UTC m=+687.687978216" watchObservedRunningTime="2025-10-10 09:22:24.676276104 +0000 UTC m=+687.692294856" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.633873 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz"] Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.635030 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.638429 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-5mrrp" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.657638 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz"] Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.661721 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp"] Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.662620 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.664365 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.691608 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp"] Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.726049 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-v5rbb"] Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.726720 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.819743 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br9tr\" (UniqueName: \"kubernetes.io/projected/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-kube-api-access-br9tr\") pod \"nmstate-webhook-6cdbc54649-dm9zp\" (UID: \"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.819814 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dm9zp\" (UID: \"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.819872 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-479dt\" (UniqueName: \"kubernetes.io/projected/35a376a3-b98d-4c9d-b7a8-633759d368ea-kube-api-access-479dt\") pod \"nmstate-metrics-fdff9cb8d-jlckz\" (UID: \"35a376a3-b98d-4c9d-b7a8-633759d368ea\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.920542 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-dbus-socket\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.920635 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-nmstate-lock\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.920701 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-ovs-socket\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.920742 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk6tb\" (UniqueName: \"kubernetes.io/projected/27433356-4791-4652-b258-aaf3d6deebef-kube-api-access-tk6tb\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.920764 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br9tr\" (UniqueName: \"kubernetes.io/projected/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-kube-api-access-br9tr\") pod \"nmstate-webhook-6cdbc54649-dm9zp\" (UID: \"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.920784 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dm9zp\" (UID: \"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.920839 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-479dt\" (UniqueName: \"kubernetes.io/projected/35a376a3-b98d-4c9d-b7a8-633759d368ea-kube-api-access-479dt\") pod \"nmstate-metrics-fdff9cb8d-jlckz\" (UID: \"35a376a3-b98d-4c9d-b7a8-633759d368ea\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" Oct 10 09:22:25 crc kubenswrapper[4669]: E1010 09:22:25.921337 4669 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 10 09:22:25 crc kubenswrapper[4669]: E1010 09:22:25.921380 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-tls-key-pair podName:9ec015c7-dd0b-46e0-a5b9-456bf97c6a49 nodeName:}" failed. No retries permitted until 2025-10-10 09:22:26.421365556 +0000 UTC m=+689.437384288 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-tls-key-pair") pod "nmstate-webhook-6cdbc54649-dm9zp" (UID: "9ec015c7-dd0b-46e0-a5b9-456bf97c6a49") : secret "openshift-nmstate-webhook" not found Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.939861 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz"] Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.940476 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.944116 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.944272 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-ns4bj" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.944368 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.951843 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz"] Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.971022 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-479dt\" (UniqueName: \"kubernetes.io/projected/35a376a3-b98d-4c9d-b7a8-633759d368ea-kube-api-access-479dt\") pod \"nmstate-metrics-fdff9cb8d-jlckz\" (UID: \"35a376a3-b98d-4c9d-b7a8-633759d368ea\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" Oct 10 09:22:25 crc kubenswrapper[4669]: I1010 09:22:25.980031 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br9tr\" (UniqueName: \"kubernetes.io/projected/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-kube-api-access-br9tr\") pod \"nmstate-webhook-6cdbc54649-dm9zp\" (UID: \"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.022091 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk6tb\" (UniqueName: \"kubernetes.io/projected/27433356-4791-4652-b258-aaf3d6deebef-kube-api-access-tk6tb\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.022184 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-dbus-socket\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.022243 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-nmstate-lock\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.022272 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-ovs-socket\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.022345 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-ovs-socket\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.023017 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-nmstate-lock\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.023638 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27433356-4791-4652-b258-aaf3d6deebef-dbus-socket\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.049361 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk6tb\" (UniqueName: \"kubernetes.io/projected/27433356-4791-4652-b258-aaf3d6deebef-kube-api-access-tk6tb\") pod \"nmstate-handler-v5rbb\" (UID: \"27433356-4791-4652-b258-aaf3d6deebef\") " pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.123414 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6d68cd46f8-6vlg9"] Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.123776 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7z4z\" (UniqueName: \"kubernetes.io/projected/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-kube-api-access-t7z4z\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.123833 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.123918 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.124125 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.142672 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d68cd46f8-6vlg9"] Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225421 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225481 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-config\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225525 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkdfz\" (UniqueName: \"kubernetes.io/projected/3ed816a9-ce76-4280-a699-aecba4bd7dff-kube-api-access-lkdfz\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225547 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-oauth-serving-cert\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225597 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-oauth-config\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225631 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-service-ca\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225668 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-serving-cert\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225703 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225724 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-trusted-ca-bundle\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.225746 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7z4z\" (UniqueName: \"kubernetes.io/projected/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-kube-api-access-t7z4z\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: E1010 09:22:26.226170 4669 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 10 09:22:26 crc kubenswrapper[4669]: E1010 09:22:26.226219 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert podName:5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0 nodeName:}" failed. No retries permitted until 2025-10-10 09:22:26.72620261 +0000 UTC m=+689.742221352 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-brznz" (UID: "5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0") : secret "plugin-serving-cert" not found Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.227521 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.242206 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7z4z\" (UniqueName: \"kubernetes.io/projected/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-kube-api-access-t7z4z\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.252977 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.326700 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-oauth-config\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.326965 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-service-ca\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.326992 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-serving-cert\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.327022 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-trusted-ca-bundle\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.327062 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-config\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.327090 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkdfz\" (UniqueName: \"kubernetes.io/projected/3ed816a9-ce76-4280-a699-aecba4bd7dff-kube-api-access-lkdfz\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.327107 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-oauth-serving-cert\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.328352 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-config\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.330709 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-oauth-config\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.331105 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-service-ca\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.331459 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ed816a9-ce76-4280-a699-aecba4bd7dff-console-serving-cert\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.331752 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-trusted-ca-bundle\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.332257 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3ed816a9-ce76-4280-a699-aecba4bd7dff-oauth-serving-cert\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.343676 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.351640 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkdfz\" (UniqueName: \"kubernetes.io/projected/3ed816a9-ce76-4280-a699-aecba4bd7dff-kube-api-access-lkdfz\") pod \"console-6d68cd46f8-6vlg9\" (UID: \"3ed816a9-ce76-4280-a699-aecba4bd7dff\") " pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.428215 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dm9zp\" (UID: \"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.433467 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9ec015c7-dd0b-46e0-a5b9-456bf97c6a49-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dm9zp\" (UID: \"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.439261 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.442321 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz"] Oct 10 09:22:26 crc kubenswrapper[4669]: W1010 09:22:26.458056 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35a376a3_b98d_4c9d_b7a8_633759d368ea.slice/crio-06729f949f0bc4197f043e380313b246a28cfd1be173ea18bf8a5522c7bf80f9 WatchSource:0}: Error finding container 06729f949f0bc4197f043e380313b246a28cfd1be173ea18bf8a5522c7bf80f9: Status 404 returned error can't find the container with id 06729f949f0bc4197f043e380313b246a28cfd1be173ea18bf8a5522c7bf80f9 Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.575876 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.636680 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6d68cd46f8-6vlg9"] Oct 10 09:22:26 crc kubenswrapper[4669]: W1010 09:22:26.645339 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ed816a9_ce76_4280_a699_aecba4bd7dff.slice/crio-613b2f512d64dec481257ea23f1e67cc3ebf2660c542c9428276cbf1d2a2d2e1 WatchSource:0}: Error finding container 613b2f512d64dec481257ea23f1e67cc3ebf2660c542c9428276cbf1d2a2d2e1: Status 404 returned error can't find the container with id 613b2f512d64dec481257ea23f1e67cc3ebf2660c542c9428276cbf1d2a2d2e1 Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.662163 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d68cd46f8-6vlg9" event={"ID":"3ed816a9-ce76-4280-a699-aecba4bd7dff","Type":"ContainerStarted","Data":"613b2f512d64dec481257ea23f1e67cc3ebf2660c542c9428276cbf1d2a2d2e1"} Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.663445 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-v5rbb" event={"ID":"27433356-4791-4652-b258-aaf3d6deebef","Type":"ContainerStarted","Data":"646e411cd502ef7c298d2b9f22c31cee77e904462b299b84b888311c9c2a503a"} Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.664596 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" event={"ID":"35a376a3-b98d-4c9d-b7a8-633759d368ea","Type":"ContainerStarted","Data":"06729f949f0bc4197f043e380313b246a28cfd1be173ea18bf8a5522c7bf80f9"} Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.733938 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:26 crc kubenswrapper[4669]: E1010 09:22:26.734087 4669 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 10 09:22:26 crc kubenswrapper[4669]: E1010 09:22:26.734148 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert podName:5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0 nodeName:}" failed. No retries permitted until 2025-10-10 09:22:27.73413525 +0000 UTC m=+690.750153992 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-brznz" (UID: "5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0") : secret "plugin-serving-cert" not found Oct 10 09:22:26 crc kubenswrapper[4669]: I1010 09:22:26.831632 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp"] Oct 10 09:22:26 crc kubenswrapper[4669]: W1010 09:22:26.838290 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ec015c7_dd0b_46e0_a5b9_456bf97c6a49.slice/crio-f6363ff9cf4a792350fa5eb23dd6963cb01a3226cdeaa8e659a8e27d2028f0ba WatchSource:0}: Error finding container f6363ff9cf4a792350fa5eb23dd6963cb01a3226cdeaa8e659a8e27d2028f0ba: Status 404 returned error can't find the container with id f6363ff9cf4a792350fa5eb23dd6963cb01a3226cdeaa8e659a8e27d2028f0ba Oct 10 09:22:27 crc kubenswrapper[4669]: I1010 09:22:27.672780 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6d68cd46f8-6vlg9" event={"ID":"3ed816a9-ce76-4280-a699-aecba4bd7dff","Type":"ContainerStarted","Data":"5a306b827cf5342e56b0dc29e4d4e9d8fbda283d22c3f7ea8ce230583e2e68dd"} Oct 10 09:22:27 crc kubenswrapper[4669]: I1010 09:22:27.675349 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" event={"ID":"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49","Type":"ContainerStarted","Data":"f6363ff9cf4a792350fa5eb23dd6963cb01a3226cdeaa8e659a8e27d2028f0ba"} Oct 10 09:22:27 crc kubenswrapper[4669]: I1010 09:22:27.696168 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6d68cd46f8-6vlg9" podStartSLOduration=1.696148092 podStartE2EDuration="1.696148092s" podCreationTimestamp="2025-10-10 09:22:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:22:27.692625428 +0000 UTC m=+690.708644170" watchObservedRunningTime="2025-10-10 09:22:27.696148092 +0000 UTC m=+690.712166834" Oct 10 09:22:27 crc kubenswrapper[4669]: I1010 09:22:27.743228 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:27 crc kubenswrapper[4669]: I1010 09:22:27.749622 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-brznz\" (UID: \"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:27 crc kubenswrapper[4669]: I1010 09:22:27.754955 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" Oct 10 09:22:27 crc kubenswrapper[4669]: I1010 09:22:27.968761 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz"] Oct 10 09:22:28 crc kubenswrapper[4669]: I1010 09:22:28.682497 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" event={"ID":"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0","Type":"ContainerStarted","Data":"4b7260e608c446f54e9da09896080ddb8f032bf487c87d9d7219223654e75469"} Oct 10 09:22:29 crc kubenswrapper[4669]: I1010 09:22:29.690842 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" event={"ID":"35a376a3-b98d-4c9d-b7a8-633759d368ea","Type":"ContainerStarted","Data":"0633d9b22edbff82b0359004cde6b5b807e0065a80de2c08d72dffd2972c515a"} Oct 10 09:22:29 crc kubenswrapper[4669]: I1010 09:22:29.693000 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" event={"ID":"9ec015c7-dd0b-46e0-a5b9-456bf97c6a49","Type":"ContainerStarted","Data":"548d2030c96d57a1ee53ed0f3d7162db65a937a0c17c40479b309a570b968efc"} Oct 10 09:22:29 crc kubenswrapper[4669]: I1010 09:22:29.693152 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:29 crc kubenswrapper[4669]: I1010 09:22:29.696027 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-v5rbb" event={"ID":"27433356-4791-4652-b258-aaf3d6deebef","Type":"ContainerStarted","Data":"92ec4a5674015bf8de77e2faa7ddc78574dba8db53b608f923daacc0b0cfc5f4"} Oct 10 09:22:29 crc kubenswrapper[4669]: I1010 09:22:29.696630 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:29 crc kubenswrapper[4669]: I1010 09:22:29.712603 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" podStartSLOduration=2.422890732 podStartE2EDuration="4.712573984s" podCreationTimestamp="2025-10-10 09:22:25 +0000 UTC" firstStartedPulling="2025-10-10 09:22:26.840142007 +0000 UTC m=+689.856160749" lastFinishedPulling="2025-10-10 09:22:29.129825259 +0000 UTC m=+692.145844001" observedRunningTime="2025-10-10 09:22:29.708082338 +0000 UTC m=+692.724101080" watchObservedRunningTime="2025-10-10 09:22:29.712573984 +0000 UTC m=+692.728592726" Oct 10 09:22:29 crc kubenswrapper[4669]: I1010 09:22:29.723765 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-v5rbb" podStartSLOduration=1.969565622 podStartE2EDuration="4.723749026s" podCreationTimestamp="2025-10-10 09:22:25 +0000 UTC" firstStartedPulling="2025-10-10 09:22:26.37672766 +0000 UTC m=+689.392746402" lastFinishedPulling="2025-10-10 09:22:29.130911064 +0000 UTC m=+692.146929806" observedRunningTime="2025-10-10 09:22:29.722044691 +0000 UTC m=+692.738063433" watchObservedRunningTime="2025-10-10 09:22:29.723749026 +0000 UTC m=+692.739767768" Oct 10 09:22:30 crc kubenswrapper[4669]: I1010 09:22:30.708665 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" event={"ID":"5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0","Type":"ContainerStarted","Data":"25bf6e2e2ff3db8d90f7d9b6e9261272cd28071a9a1306d3ec957a70b45672cc"} Oct 10 09:22:32 crc kubenswrapper[4669]: I1010 09:22:32.724056 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" event={"ID":"35a376a3-b98d-4c9d-b7a8-633759d368ea","Type":"ContainerStarted","Data":"3fa638715d979e1f83e69e808f87f3f953a0d6079db1ab40332ac240d9ecd95c"} Oct 10 09:22:32 crc kubenswrapper[4669]: I1010 09:22:32.756063 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-jlckz" podStartSLOduration=2.462275719 podStartE2EDuration="7.756027507s" podCreationTimestamp="2025-10-10 09:22:25 +0000 UTC" firstStartedPulling="2025-10-10 09:22:26.459909568 +0000 UTC m=+689.475928310" lastFinishedPulling="2025-10-10 09:22:31.753661356 +0000 UTC m=+694.769680098" observedRunningTime="2025-10-10 09:22:32.754210247 +0000 UTC m=+695.770229029" watchObservedRunningTime="2025-10-10 09:22:32.756027507 +0000 UTC m=+695.772046299" Oct 10 09:22:32 crc kubenswrapper[4669]: I1010 09:22:32.761630 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-brznz" podStartSLOduration=5.408422208 podStartE2EDuration="7.761567487s" podCreationTimestamp="2025-10-10 09:22:25 +0000 UTC" firstStartedPulling="2025-10-10 09:22:27.974911732 +0000 UTC m=+690.990930464" lastFinishedPulling="2025-10-10 09:22:30.328057001 +0000 UTC m=+693.344075743" observedRunningTime="2025-10-10 09:22:30.736900517 +0000 UTC m=+693.752919329" watchObservedRunningTime="2025-10-10 09:22:32.761567487 +0000 UTC m=+695.777586269" Oct 10 09:22:36 crc kubenswrapper[4669]: I1010 09:22:36.373820 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-v5rbb" Oct 10 09:22:36 crc kubenswrapper[4669]: I1010 09:22:36.439739 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:36 crc kubenswrapper[4669]: I1010 09:22:36.439864 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:36 crc kubenswrapper[4669]: I1010 09:22:36.445616 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:36 crc kubenswrapper[4669]: I1010 09:22:36.752276 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6d68cd46f8-6vlg9" Oct 10 09:22:36 crc kubenswrapper[4669]: I1010 09:22:36.829346 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wqc7t"] Oct 10 09:22:46 crc kubenswrapper[4669]: I1010 09:22:46.584654 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dm9zp" Oct 10 09:22:54 crc kubenswrapper[4669]: I1010 09:22:54.274999 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:22:54 crc kubenswrapper[4669]: I1010 09:22:54.275650 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.051976 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658"] Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.053412 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.063877 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658"] Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.064749 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.174473 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.174542 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.174768 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dzj2\" (UniqueName: \"kubernetes.io/projected/5e1484dd-9b20-445b-9aaa-30884d9fd717-kube-api-access-4dzj2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.276151 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.276306 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dzj2\" (UniqueName: \"kubernetes.io/projected/5e1484dd-9b20-445b-9aaa-30884d9fd717-kube-api-access-4dzj2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.276389 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.276894 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.277087 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.296333 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dzj2\" (UniqueName: \"kubernetes.io/projected/5e1484dd-9b20-445b-9aaa-30884d9fd717-kube-api-access-4dzj2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.374805 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.552907 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658"] Oct 10 09:23:00 crc kubenswrapper[4669]: W1010 09:23:00.558109 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e1484dd_9b20_445b_9aaa_30884d9fd717.slice/crio-0ab0800edbaee877e30546d9c5747e98dc8b1a8294a2b370855931f2ef3d4fde WatchSource:0}: Error finding container 0ab0800edbaee877e30546d9c5747e98dc8b1a8294a2b370855931f2ef3d4fde: Status 404 returned error can't find the container with id 0ab0800edbaee877e30546d9c5747e98dc8b1a8294a2b370855931f2ef3d4fde Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.891397 4669 generic.go:334] "Generic (PLEG): container finished" podID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerID="4f102e3871df65c322e0ec494fe360e3590ca636a0ca1ce5472a5fd650817e17" exitCode=0 Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.891689 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" event={"ID":"5e1484dd-9b20-445b-9aaa-30884d9fd717","Type":"ContainerDied","Data":"4f102e3871df65c322e0ec494fe360e3590ca636a0ca1ce5472a5fd650817e17"} Oct 10 09:23:00 crc kubenswrapper[4669]: I1010 09:23:00.891813 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" event={"ID":"5e1484dd-9b20-445b-9aaa-30884d9fd717","Type":"ContainerStarted","Data":"0ab0800edbaee877e30546d9c5747e98dc8b1a8294a2b370855931f2ef3d4fde"} Oct 10 09:23:01 crc kubenswrapper[4669]: I1010 09:23:01.886002 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-wqc7t" podUID="69730259-7614-4869-bc84-10e17d8d13df" containerName="console" containerID="cri-o://e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b" gracePeriod=15 Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.198908 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wqc7t_69730259-7614-4869-bc84-10e17d8d13df/console/0.log" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.199092 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.304837 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-oauth-config\") pod \"69730259-7614-4869-bc84-10e17d8d13df\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.304901 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shld6\" (UniqueName: \"kubernetes.io/projected/69730259-7614-4869-bc84-10e17d8d13df-kube-api-access-shld6\") pod \"69730259-7614-4869-bc84-10e17d8d13df\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.304940 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-console-config\") pod \"69730259-7614-4869-bc84-10e17d8d13df\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.304990 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-trusted-ca-bundle\") pod \"69730259-7614-4869-bc84-10e17d8d13df\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.305046 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-service-ca\") pod \"69730259-7614-4869-bc84-10e17d8d13df\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.305118 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-oauth-serving-cert\") pod \"69730259-7614-4869-bc84-10e17d8d13df\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.305173 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-serving-cert\") pod \"69730259-7614-4869-bc84-10e17d8d13df\" (UID: \"69730259-7614-4869-bc84-10e17d8d13df\") " Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.305737 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-console-config" (OuterVolumeSpecName: "console-config") pod "69730259-7614-4869-bc84-10e17d8d13df" (UID: "69730259-7614-4869-bc84-10e17d8d13df"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.305783 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "69730259-7614-4869-bc84-10e17d8d13df" (UID: "69730259-7614-4869-bc84-10e17d8d13df"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.306186 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-service-ca" (OuterVolumeSpecName: "service-ca") pod "69730259-7614-4869-bc84-10e17d8d13df" (UID: "69730259-7614-4869-bc84-10e17d8d13df"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.306250 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "69730259-7614-4869-bc84-10e17d8d13df" (UID: "69730259-7614-4869-bc84-10e17d8d13df"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.317183 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "69730259-7614-4869-bc84-10e17d8d13df" (UID: "69730259-7614-4869-bc84-10e17d8d13df"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.317291 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69730259-7614-4869-bc84-10e17d8d13df-kube-api-access-shld6" (OuterVolumeSpecName: "kube-api-access-shld6") pod "69730259-7614-4869-bc84-10e17d8d13df" (UID: "69730259-7614-4869-bc84-10e17d8d13df"). InnerVolumeSpecName "kube-api-access-shld6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.317291 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "69730259-7614-4869-bc84-10e17d8d13df" (UID: "69730259-7614-4869-bc84-10e17d8d13df"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.408529 4669 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-console-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.408570 4669 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.408606 4669 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.408621 4669 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/69730259-7614-4869-bc84-10e17d8d13df-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.408633 4669 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.408645 4669 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/69730259-7614-4869-bc84-10e17d8d13df-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.408655 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shld6\" (UniqueName: \"kubernetes.io/projected/69730259-7614-4869-bc84-10e17d8d13df-kube-api-access-shld6\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.906528 4669 generic.go:334] "Generic (PLEG): container finished" podID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerID="1903eb7edf015c8f6cdf5cdd0b23074985322a99c86127cd0ba43b949d12f1a9" exitCode=0 Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.906748 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" event={"ID":"5e1484dd-9b20-445b-9aaa-30884d9fd717","Type":"ContainerDied","Data":"1903eb7edf015c8f6cdf5cdd0b23074985322a99c86127cd0ba43b949d12f1a9"} Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.910973 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wqc7t_69730259-7614-4869-bc84-10e17d8d13df/console/0.log" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.911013 4669 generic.go:334] "Generic (PLEG): container finished" podID="69730259-7614-4869-bc84-10e17d8d13df" containerID="e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b" exitCode=2 Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.911069 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wqc7t" event={"ID":"69730259-7614-4869-bc84-10e17d8d13df","Type":"ContainerDied","Data":"e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b"} Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.911094 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wqc7t" event={"ID":"69730259-7614-4869-bc84-10e17d8d13df","Type":"ContainerDied","Data":"d7c3e8e876a4fd623fab152ba220421901928f6a9517bbef18f7ecf4013de2a2"} Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.911141 4669 scope.go:117] "RemoveContainer" containerID="e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.911312 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wqc7t" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.942461 4669 scope.go:117] "RemoveContainer" containerID="e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b" Oct 10 09:23:02 crc kubenswrapper[4669]: E1010 09:23:02.942949 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b\": container with ID starting with e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b not found: ID does not exist" containerID="e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.942976 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b"} err="failed to get container status \"e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b\": rpc error: code = NotFound desc = could not find container \"e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b\": container with ID starting with e9689549ab4adf2adad935cd315e8bd57a0dd6cff443450b3cb52490bb7d9a7b not found: ID does not exist" Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.960271 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wqc7t"] Oct 10 09:23:02 crc kubenswrapper[4669]: I1010 09:23:02.964242 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-wqc7t"] Oct 10 09:23:03 crc kubenswrapper[4669]: I1010 09:23:03.803941 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69730259-7614-4869-bc84-10e17d8d13df" path="/var/lib/kubelet/pods/69730259-7614-4869-bc84-10e17d8d13df/volumes" Oct 10 09:23:03 crc kubenswrapper[4669]: I1010 09:23:03.924173 4669 generic.go:334] "Generic (PLEG): container finished" podID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerID="43ec09b69e35f2d97abac3cb3dbcdb18a5070849d9f9ce5ac92df9e0ee83d53c" exitCode=0 Oct 10 09:23:03 crc kubenswrapper[4669]: I1010 09:23:03.924281 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" event={"ID":"5e1484dd-9b20-445b-9aaa-30884d9fd717","Type":"ContainerDied","Data":"43ec09b69e35f2d97abac3cb3dbcdb18a5070849d9f9ce5ac92df9e0ee83d53c"} Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.152634 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.253217 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dzj2\" (UniqueName: \"kubernetes.io/projected/5e1484dd-9b20-445b-9aaa-30884d9fd717-kube-api-access-4dzj2\") pod \"5e1484dd-9b20-445b-9aaa-30884d9fd717\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.253267 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-bundle\") pod \"5e1484dd-9b20-445b-9aaa-30884d9fd717\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.253312 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-util\") pod \"5e1484dd-9b20-445b-9aaa-30884d9fd717\" (UID: \"5e1484dd-9b20-445b-9aaa-30884d9fd717\") " Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.255486 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-bundle" (OuterVolumeSpecName: "bundle") pod "5e1484dd-9b20-445b-9aaa-30884d9fd717" (UID: "5e1484dd-9b20-445b-9aaa-30884d9fd717"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.262285 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1484dd-9b20-445b-9aaa-30884d9fd717-kube-api-access-4dzj2" (OuterVolumeSpecName: "kube-api-access-4dzj2") pod "5e1484dd-9b20-445b-9aaa-30884d9fd717" (UID: "5e1484dd-9b20-445b-9aaa-30884d9fd717"). InnerVolumeSpecName "kube-api-access-4dzj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.266964 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-util" (OuterVolumeSpecName: "util") pod "5e1484dd-9b20-445b-9aaa-30884d9fd717" (UID: "5e1484dd-9b20-445b-9aaa-30884d9fd717"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.354952 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dzj2\" (UniqueName: \"kubernetes.io/projected/5e1484dd-9b20-445b-9aaa-30884d9fd717-kube-api-access-4dzj2\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.354983 4669 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.354992 4669 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5e1484dd-9b20-445b-9aaa-30884d9fd717-util\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.953077 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" event={"ID":"5e1484dd-9b20-445b-9aaa-30884d9fd717","Type":"ContainerDied","Data":"0ab0800edbaee877e30546d9c5747e98dc8b1a8294a2b370855931f2ef3d4fde"} Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.953121 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ab0800edbaee877e30546d9c5747e98dc8b1a8294a2b370855931f2ef3d4fde" Oct 10 09:23:05 crc kubenswrapper[4669]: I1010 09:23:05.953207 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.763997 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4"] Oct 10 09:23:13 crc kubenswrapper[4669]: E1010 09:23:13.764799 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerName="util" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.764816 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerName="util" Oct 10 09:23:13 crc kubenswrapper[4669]: E1010 09:23:13.764829 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerName="extract" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.764837 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerName="extract" Oct 10 09:23:13 crc kubenswrapper[4669]: E1010 09:23:13.764865 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerName="pull" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.764873 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerName="pull" Oct 10 09:23:13 crc kubenswrapper[4669]: E1010 09:23:13.764887 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69730259-7614-4869-bc84-10e17d8d13df" containerName="console" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.764895 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="69730259-7614-4869-bc84-10e17d8d13df" containerName="console" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.765020 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e1484dd-9b20-445b-9aaa-30884d9fd717" containerName="extract" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.765036 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="69730259-7614-4869-bc84-10e17d8d13df" containerName="console" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.765532 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.784419 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.784775 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.786564 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-6b456" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.787294 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.787738 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.811479 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4"] Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.865110 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggk4t\" (UniqueName: \"kubernetes.io/projected/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-kube-api-access-ggk4t\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.865157 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-apiservice-cert\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.865425 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-webhook-cert\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.966931 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-webhook-cert\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.967247 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggk4t\" (UniqueName: \"kubernetes.io/projected/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-kube-api-access-ggk4t\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.967272 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-apiservice-cert\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.971730 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-webhook-cert\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:13 crc kubenswrapper[4669]: I1010 09:23:13.991782 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-apiservice-cert\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.009012 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggk4t\" (UniqueName: \"kubernetes.io/projected/5cda559e-d9fe-4651-b77d-d8c7d38e6a21-kube-api-access-ggk4t\") pod \"metallb-operator-controller-manager-79d96cb9f5-vkfb4\" (UID: \"5cda559e-d9fe-4651-b77d-d8c7d38e6a21\") " pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.082687 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.120411 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42"] Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.121533 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.123851 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5ng9d" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.123994 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.125069 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.139906 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42"] Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.274293 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06827887-c1ea-4839-9ed3-a4c8813d81ac-webhook-cert\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.274661 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntftx\" (UniqueName: \"kubernetes.io/projected/06827887-c1ea-4839-9ed3-a4c8813d81ac-kube-api-access-ntftx\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.274720 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06827887-c1ea-4839-9ed3-a4c8813d81ac-apiservice-cert\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.379314 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06827887-c1ea-4839-9ed3-a4c8813d81ac-apiservice-cert\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.379388 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06827887-c1ea-4839-9ed3-a4c8813d81ac-webhook-cert\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.379459 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntftx\" (UniqueName: \"kubernetes.io/projected/06827887-c1ea-4839-9ed3-a4c8813d81ac-kube-api-access-ntftx\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.384985 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06827887-c1ea-4839-9ed3-a4c8813d81ac-apiservice-cert\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.388362 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06827887-c1ea-4839-9ed3-a4c8813d81ac-webhook-cert\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.410186 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4"] Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.419780 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntftx\" (UniqueName: \"kubernetes.io/projected/06827887-c1ea-4839-9ed3-a4c8813d81ac-kube-api-access-ntftx\") pod \"metallb-operator-webhook-server-7dd78c579-pxx42\" (UID: \"06827887-c1ea-4839-9ed3-a4c8813d81ac\") " pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: W1010 09:23:14.421072 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cda559e_d9fe_4651_b77d_d8c7d38e6a21.slice/crio-b25f80712eb499f1321bc7d27c3cbc5ed0754049e528892fa783543899554b52 WatchSource:0}: Error finding container b25f80712eb499f1321bc7d27c3cbc5ed0754049e528892fa783543899554b52: Status 404 returned error can't find the container with id b25f80712eb499f1321bc7d27c3cbc5ed0754049e528892fa783543899554b52 Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.463852 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:14 crc kubenswrapper[4669]: I1010 09:23:14.877126 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42"] Oct 10 09:23:14 crc kubenswrapper[4669]: W1010 09:23:14.887577 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06827887_c1ea_4839_9ed3_a4c8813d81ac.slice/crio-b90792c74d75869edb6b84b91cdf6dbbc027241ac500917c7aff2467ec23a27d WatchSource:0}: Error finding container b90792c74d75869edb6b84b91cdf6dbbc027241ac500917c7aff2467ec23a27d: Status 404 returned error can't find the container with id b90792c74d75869edb6b84b91cdf6dbbc027241ac500917c7aff2467ec23a27d Oct 10 09:23:15 crc kubenswrapper[4669]: I1010 09:23:15.002267 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" event={"ID":"06827887-c1ea-4839-9ed3-a4c8813d81ac","Type":"ContainerStarted","Data":"b90792c74d75869edb6b84b91cdf6dbbc027241ac500917c7aff2467ec23a27d"} Oct 10 09:23:15 crc kubenswrapper[4669]: I1010 09:23:15.003261 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" event={"ID":"5cda559e-d9fe-4651-b77d-d8c7d38e6a21","Type":"ContainerStarted","Data":"b25f80712eb499f1321bc7d27c3cbc5ed0754049e528892fa783543899554b52"} Oct 10 09:23:21 crc kubenswrapper[4669]: I1010 09:23:21.037814 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" event={"ID":"5cda559e-d9fe-4651-b77d-d8c7d38e6a21","Type":"ContainerStarted","Data":"c62d705f23eb3bad291c2354abeac6dd615b260f5f7d463cfe99c3b93da068b8"} Oct 10 09:23:21 crc kubenswrapper[4669]: I1010 09:23:21.039174 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:21 crc kubenswrapper[4669]: I1010 09:23:21.040641 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" event={"ID":"06827887-c1ea-4839-9ed3-a4c8813d81ac","Type":"ContainerStarted","Data":"aca0dc0c010ade0358411e2eb402c0f74db7157f258ccbc14680436d91f325b4"} Oct 10 09:23:21 crc kubenswrapper[4669]: I1010 09:23:21.041004 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:21 crc kubenswrapper[4669]: I1010 09:23:21.059526 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" podStartSLOduration=1.936239775 podStartE2EDuration="8.059510069s" podCreationTimestamp="2025-10-10 09:23:13 +0000 UTC" firstStartedPulling="2025-10-10 09:23:14.424496158 +0000 UTC m=+737.440514900" lastFinishedPulling="2025-10-10 09:23:20.547766442 +0000 UTC m=+743.563785194" observedRunningTime="2025-10-10 09:23:21.057712561 +0000 UTC m=+744.073731303" watchObservedRunningTime="2025-10-10 09:23:21.059510069 +0000 UTC m=+744.075528801" Oct 10 09:23:21 crc kubenswrapper[4669]: I1010 09:23:21.081809 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" podStartSLOduration=1.383595465 podStartE2EDuration="7.081789851s" podCreationTimestamp="2025-10-10 09:23:14 +0000 UTC" firstStartedPulling="2025-10-10 09:23:14.89043969 +0000 UTC m=+737.906458432" lastFinishedPulling="2025-10-10 09:23:20.588634066 +0000 UTC m=+743.604652818" observedRunningTime="2025-10-10 09:23:21.077195812 +0000 UTC m=+744.093214554" watchObservedRunningTime="2025-10-10 09:23:21.081789851 +0000 UTC m=+744.097808593" Oct 10 09:23:24 crc kubenswrapper[4669]: I1010 09:23:24.274401 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:23:24 crc kubenswrapper[4669]: I1010 09:23:24.274898 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.410296 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8q8w9"] Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.410964 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" podUID="0342148d-ffe9-4d1d-a629-03108f4c2cda" containerName="controller-manager" containerID="cri-o://f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739" gracePeriod=30 Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.504747 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg"] Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.505032 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" podUID="2a92e481-9a5c-4238-988f-f2d0b852ebca" containerName="route-controller-manager" containerID="cri-o://8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f" gracePeriod=30 Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.846924 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.947384 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-client-ca\") pod \"0342148d-ffe9-4d1d-a629-03108f4c2cda\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.947461 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0342148d-ffe9-4d1d-a629-03108f4c2cda-serving-cert\") pod \"0342148d-ffe9-4d1d-a629-03108f4c2cda\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.947504 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-config\") pod \"0342148d-ffe9-4d1d-a629-03108f4c2cda\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.947531 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq9qk\" (UniqueName: \"kubernetes.io/projected/0342148d-ffe9-4d1d-a629-03108f4c2cda-kube-api-access-kq9qk\") pod \"0342148d-ffe9-4d1d-a629-03108f4c2cda\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.947547 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-proxy-ca-bundles\") pod \"0342148d-ffe9-4d1d-a629-03108f4c2cda\" (UID: \"0342148d-ffe9-4d1d-a629-03108f4c2cda\") " Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.948401 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0342148d-ffe9-4d1d-a629-03108f4c2cda" (UID: "0342148d-ffe9-4d1d-a629-03108f4c2cda"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.948391 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-client-ca" (OuterVolumeSpecName: "client-ca") pod "0342148d-ffe9-4d1d-a629-03108f4c2cda" (UID: "0342148d-ffe9-4d1d-a629-03108f4c2cda"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.948705 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-config" (OuterVolumeSpecName: "config") pod "0342148d-ffe9-4d1d-a629-03108f4c2cda" (UID: "0342148d-ffe9-4d1d-a629-03108f4c2cda"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.953682 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0342148d-ffe9-4d1d-a629-03108f4c2cda-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0342148d-ffe9-4d1d-a629-03108f4c2cda" (UID: "0342148d-ffe9-4d1d-a629-03108f4c2cda"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.956183 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0342148d-ffe9-4d1d-a629-03108f4c2cda-kube-api-access-kq9qk" (OuterVolumeSpecName: "kube-api-access-kq9qk") pod "0342148d-ffe9-4d1d-a629-03108f4c2cda" (UID: "0342148d-ffe9-4d1d-a629-03108f4c2cda"). InnerVolumeSpecName "kube-api-access-kq9qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:23:26 crc kubenswrapper[4669]: I1010 09:23:26.988032 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.048898 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.048940 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq9qk\" (UniqueName: \"kubernetes.io/projected/0342148d-ffe9-4d1d-a629-03108f4c2cda-kube-api-access-kq9qk\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.048958 4669 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.048970 4669 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0342148d-ffe9-4d1d-a629-03108f4c2cda-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.048983 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0342148d-ffe9-4d1d-a629-03108f4c2cda-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.074827 4669 generic.go:334] "Generic (PLEG): container finished" podID="0342148d-ffe9-4d1d-a629-03108f4c2cda" containerID="f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739" exitCode=0 Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.074881 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" event={"ID":"0342148d-ffe9-4d1d-a629-03108f4c2cda","Type":"ContainerDied","Data":"f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739"} Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.074900 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.074945 4669 scope.go:117] "RemoveContainer" containerID="f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.074930 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8q8w9" event={"ID":"0342148d-ffe9-4d1d-a629-03108f4c2cda","Type":"ContainerDied","Data":"13102c2875d2040863f0c3bb8410ca94fb7823f14721508cf3794edb55e959e7"} Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.076576 4669 generic.go:334] "Generic (PLEG): container finished" podID="2a92e481-9a5c-4238-988f-f2d0b852ebca" containerID="8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f" exitCode=0 Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.076625 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" event={"ID":"2a92e481-9a5c-4238-988f-f2d0b852ebca","Type":"ContainerDied","Data":"8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f"} Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.076650 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" event={"ID":"2a92e481-9a5c-4238-988f-f2d0b852ebca","Type":"ContainerDied","Data":"6f13ad89718ae3f7c0a8663fcfa00b19f084320ef2b57596a0462a857ac5faf0"} Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.076659 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.092141 4669 scope.go:117] "RemoveContainer" containerID="f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739" Oct 10 09:23:27 crc kubenswrapper[4669]: E1010 09:23:27.092679 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739\": container with ID starting with f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739 not found: ID does not exist" containerID="f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.092717 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739"} err="failed to get container status \"f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739\": rpc error: code = NotFound desc = could not find container \"f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739\": container with ID starting with f53e4ea260ace61cd88867c154af8bcd03bc481c080f941e482084a0d381d739 not found: ID does not exist" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.092752 4669 scope.go:117] "RemoveContainer" containerID="8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.115447 4669 scope.go:117] "RemoveContainer" containerID="8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f" Oct 10 09:23:27 crc kubenswrapper[4669]: E1010 09:23:27.116058 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f\": container with ID starting with 8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f not found: ID does not exist" containerID="8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.116176 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f"} err="failed to get container status \"8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f\": rpc error: code = NotFound desc = could not find container \"8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f\": container with ID starting with 8fce78ffc3f1427f6c6e5f7e074fb861a96ba4c0ac33982ae0584d9a2a24b61f not found: ID does not exist" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.124236 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8q8w9"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.128043 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8q8w9"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.150318 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config\") pod \"2a92e481-9a5c-4238-988f-f2d0b852ebca\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.150548 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert\") pod \"2a92e481-9a5c-4238-988f-f2d0b852ebca\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.150672 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca\") pod \"2a92e481-9a5c-4238-988f-f2d0b852ebca\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.150839 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggp22\" (UniqueName: \"kubernetes.io/projected/2a92e481-9a5c-4238-988f-f2d0b852ebca-kube-api-access-ggp22\") pod \"2a92e481-9a5c-4238-988f-f2d0b852ebca\" (UID: \"2a92e481-9a5c-4238-988f-f2d0b852ebca\") " Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.151164 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config" (OuterVolumeSpecName: "config") pod "2a92e481-9a5c-4238-988f-f2d0b852ebca" (UID: "2a92e481-9a5c-4238-988f-f2d0b852ebca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.151193 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca" (OuterVolumeSpecName: "client-ca") pod "2a92e481-9a5c-4238-988f-f2d0b852ebca" (UID: "2a92e481-9a5c-4238-988f-f2d0b852ebca"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.153933 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a92e481-9a5c-4238-988f-f2d0b852ebca-kube-api-access-ggp22" (OuterVolumeSpecName: "kube-api-access-ggp22") pod "2a92e481-9a5c-4238-988f-f2d0b852ebca" (UID: "2a92e481-9a5c-4238-988f-f2d0b852ebca"). InnerVolumeSpecName "kube-api-access-ggp22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.153995 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2a92e481-9a5c-4238-988f-f2d0b852ebca" (UID: "2a92e481-9a5c-4238-988f-f2d0b852ebca"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.252220 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.252256 4669 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a92e481-9a5c-4238-988f-f2d0b852ebca-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.252266 4669 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a92e481-9a5c-4238-988f-f2d0b852ebca-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.252274 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggp22\" (UniqueName: \"kubernetes.io/projected/2a92e481-9a5c-4238-988f-f2d0b852ebca-kube-api-access-ggp22\") on node \"crc\" DevicePath \"\"" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.297729 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-59ff448dc7-7n5s4"] Oct 10 09:23:27 crc kubenswrapper[4669]: E1010 09:23:27.297950 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0342148d-ffe9-4d1d-a629-03108f4c2cda" containerName="controller-manager" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.297965 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0342148d-ffe9-4d1d-a629-03108f4c2cda" containerName="controller-manager" Oct 10 09:23:27 crc kubenswrapper[4669]: E1010 09:23:27.297984 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a92e481-9a5c-4238-988f-f2d0b852ebca" containerName="route-controller-manager" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.297990 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a92e481-9a5c-4238-988f-f2d0b852ebca" containerName="route-controller-manager" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.298103 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a92e481-9a5c-4238-988f-f2d0b852ebca" containerName="route-controller-manager" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.298116 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0342148d-ffe9-4d1d-a629-03108f4c2cda" containerName="controller-manager" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.298526 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.301759 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.302024 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.305289 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.305303 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.305289 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.305303 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.319203 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59ff448dc7-7n5s4"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.324521 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.344156 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.345008 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.402412 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455239 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-config\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455310 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f74d3702-6e4c-41c9-bcb8-10e212612020-serving-cert\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455332 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g4df\" (UniqueName: \"kubernetes.io/projected/f74d3702-6e4c-41c9-bcb8-10e212612020-kube-api-access-4g4df\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455365 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-serving-cert\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455398 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74d3702-6e4c-41c9-bcb8-10e212612020-config\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455412 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-client-ca\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455441 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nlxb\" (UniqueName: \"kubernetes.io/projected/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-kube-api-access-6nlxb\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455470 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f74d3702-6e4c-41c9-bcb8-10e212612020-client-ca\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.455494 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-proxy-ca-bundles\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.456175 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.464566 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l67bg"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556186 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f74d3702-6e4c-41c9-bcb8-10e212612020-client-ca\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556224 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nlxb\" (UniqueName: \"kubernetes.io/projected/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-kube-api-access-6nlxb\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556251 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-proxy-ca-bundles\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556292 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-config\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556328 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f74d3702-6e4c-41c9-bcb8-10e212612020-serving-cert\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556347 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g4df\" (UniqueName: \"kubernetes.io/projected/f74d3702-6e4c-41c9-bcb8-10e212612020-kube-api-access-4g4df\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556376 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-client-ca\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556391 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-serving-cert\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.556409 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74d3702-6e4c-41c9-bcb8-10e212612020-config\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.557789 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f74d3702-6e4c-41c9-bcb8-10e212612020-client-ca\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.557796 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74d3702-6e4c-41c9-bcb8-10e212612020-config\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.558182 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-proxy-ca-bundles\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.559047 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-client-ca\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.560698 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-config\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.570533 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-serving-cert\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.570567 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f74d3702-6e4c-41c9-bcb8-10e212612020-serving-cert\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.588359 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nlxb\" (UniqueName: \"kubernetes.io/projected/5ba2cf49-d4d9-4b29-9bec-4476c5e23723-kube-api-access-6nlxb\") pod \"controller-manager-59ff448dc7-7n5s4\" (UID: \"5ba2cf49-d4d9-4b29-9bec-4476c5e23723\") " pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.592486 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g4df\" (UniqueName: \"kubernetes.io/projected/f74d3702-6e4c-41c9-bcb8-10e212612020-kube-api-access-4g4df\") pod \"route-controller-manager-7845959769-gc4w7\" (UID: \"f74d3702-6e4c-41c9-bcb8-10e212612020\") " pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.612261 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.665836 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.813235 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0342148d-ffe9-4d1d-a629-03108f4c2cda" path="/var/lib/kubelet/pods/0342148d-ffe9-4d1d-a629-03108f4c2cda/volumes" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.813889 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a92e481-9a5c-4238-988f-f2d0b852ebca" path="/var/lib/kubelet/pods/2a92e481-9a5c-4238-988f-f2d0b852ebca/volumes" Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.898947 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59ff448dc7-7n5s4"] Oct 10 09:23:27 crc kubenswrapper[4669]: I1010 09:23:27.997465 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7"] Oct 10 09:23:28 crc kubenswrapper[4669]: I1010 09:23:28.082930 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" event={"ID":"f74d3702-6e4c-41c9-bcb8-10e212612020","Type":"ContainerStarted","Data":"e98843cb45851a0b103b7064fae7af04afed2902a745a537b4591e8b9a75086d"} Oct 10 09:23:28 crc kubenswrapper[4669]: I1010 09:23:28.084536 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" event={"ID":"5ba2cf49-d4d9-4b29-9bec-4476c5e23723","Type":"ContainerStarted","Data":"043b6a141a9b28ad5ced96d99d458500ff8367c97098be77daee7f4db93746a9"} Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.092433 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" event={"ID":"f74d3702-6e4c-41c9-bcb8-10e212612020","Type":"ContainerStarted","Data":"93fa957bfb811c747196a2748d9f453b752890cd309df8bb9375b3614db0a40d"} Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.092718 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.094197 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" event={"ID":"5ba2cf49-d4d9-4b29-9bec-4476c5e23723","Type":"ContainerStarted","Data":"f51b65f87bf85c2a2b440edf3e32a0277d74217ccf005e42c8ab2aab6c13a3eb"} Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.095026 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.098182 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.101463 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.128475 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7845959769-gc4w7" podStartSLOduration=2.128460726 podStartE2EDuration="2.128460726s" podCreationTimestamp="2025-10-10 09:23:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:23:29.127504085 +0000 UTC m=+752.143522847" watchObservedRunningTime="2025-10-10 09:23:29.128460726 +0000 UTC m=+752.144479468" Oct 10 09:23:29 crc kubenswrapper[4669]: I1010 09:23:29.159743 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-59ff448dc7-7n5s4" podStartSLOduration=2.15972677 podStartE2EDuration="2.15972677s" podCreationTimestamp="2025-10-10 09:23:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:23:29.155936047 +0000 UTC m=+752.171954789" watchObservedRunningTime="2025-10-10 09:23:29.15972677 +0000 UTC m=+752.175745512" Oct 10 09:23:30 crc kubenswrapper[4669]: I1010 09:23:30.723108 4669 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 09:23:34 crc kubenswrapper[4669]: I1010 09:23:34.478371 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7dd78c579-pxx42" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.087332 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-79d96cb9f5-vkfb4" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.275601 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.275700 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.275770 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.276802 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1be725249e1775acac8f350a1439ffc4d89a22bc933af6755f8ceefcd7eb018"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.276880 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://b1be725249e1775acac8f350a1439ffc4d89a22bc933af6755f8ceefcd7eb018" gracePeriod=600 Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.875662 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll"] Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.877046 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.881541 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.886334 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-rp4mn"] Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.886753 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-74px6" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.889016 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.891249 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.895169 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll"] Oct 10 09:23:54 crc kubenswrapper[4669]: I1010 09:23:54.896733 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.011149 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-sockets\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.011424 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-conf\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.011664 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z6c6\" (UniqueName: \"kubernetes.io/projected/8bd73190-2f9a-45df-8e2d-8579238bf4a3-kube-api-access-4z6c6\") pod \"frr-k8s-webhook-server-64bf5d555-rgcll\" (UID: \"8bd73190-2f9a-45df-8e2d-8579238bf4a3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.011736 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-metrics\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.011895 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdqxf\" (UniqueName: \"kubernetes.io/projected/3f8d268c-022e-42b5-8e08-dbdefd895e23-kube-api-access-hdqxf\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.011970 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-startup\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.012024 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f8d268c-022e-42b5-8e08-dbdefd895e23-metrics-certs\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.012056 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8bd73190-2f9a-45df-8e2d-8579238bf4a3-cert\") pod \"frr-k8s-webhook-server-64bf5d555-rgcll\" (UID: \"8bd73190-2f9a-45df-8e2d-8579238bf4a3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.012243 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-reloader\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.025343 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-78hzw"] Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.026434 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.029539 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.029670 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.031227 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bjbjc" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.032111 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.037083 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-xh4cl"] Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.037961 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.039444 4669 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.067435 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-xh4cl"] Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.113404 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-startup\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114530 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmwrm\" (UniqueName: \"kubernetes.io/projected/ec1f1180-0070-43bd-9ee3-095e8e2379f0-kube-api-access-nmwrm\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114653 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-cert\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114793 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f8d268c-022e-42b5-8e08-dbdefd895e23-metrics-certs\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114829 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8bd73190-2f9a-45df-8e2d-8579238bf4a3-cert\") pod \"frr-k8s-webhook-server-64bf5d555-rgcll\" (UID: \"8bd73190-2f9a-45df-8e2d-8579238bf4a3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114864 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metrics-certs\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114884 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114910 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-reloader\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114940 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-sockets\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114955 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-conf\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.114990 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-metrics-certs\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115011 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z6c6\" (UniqueName: \"kubernetes.io/projected/8bd73190-2f9a-45df-8e2d-8579238bf4a3-kube-api-access-4z6c6\") pod \"frr-k8s-webhook-server-64bf5d555-rgcll\" (UID: \"8bd73190-2f9a-45df-8e2d-8579238bf4a3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115026 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4slzs\" (UniqueName: \"kubernetes.io/projected/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-kube-api-access-4slzs\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115043 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-metrics\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115067 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metallb-excludel2\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115090 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdqxf\" (UniqueName: \"kubernetes.io/projected/3f8d268c-022e-42b5-8e08-dbdefd895e23-kube-api-access-hdqxf\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115152 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-startup\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115393 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-sockets\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115783 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-metrics\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.115823 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-frr-conf\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.116059 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3f8d268c-022e-42b5-8e08-dbdefd895e23-reloader\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.125332 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8bd73190-2f9a-45df-8e2d-8579238bf4a3-cert\") pod \"frr-k8s-webhook-server-64bf5d555-rgcll\" (UID: \"8bd73190-2f9a-45df-8e2d-8579238bf4a3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.135768 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f8d268c-022e-42b5-8e08-dbdefd895e23-metrics-certs\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.138695 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z6c6\" (UniqueName: \"kubernetes.io/projected/8bd73190-2f9a-45df-8e2d-8579238bf4a3-kube-api-access-4z6c6\") pod \"frr-k8s-webhook-server-64bf5d555-rgcll\" (UID: \"8bd73190-2f9a-45df-8e2d-8579238bf4a3\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.142903 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdqxf\" (UniqueName: \"kubernetes.io/projected/3f8d268c-022e-42b5-8e08-dbdefd895e23-kube-api-access-hdqxf\") pod \"frr-k8s-rp4mn\" (UID: \"3f8d268c-022e-42b5-8e08-dbdefd895e23\") " pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.190457 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.205534 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.216236 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-metrics-certs\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.216298 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4slzs\" (UniqueName: \"kubernetes.io/projected/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-kube-api-access-4slzs\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.216336 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metallb-excludel2\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.216384 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-cert\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.216405 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmwrm\" (UniqueName: \"kubernetes.io/projected/ec1f1180-0070-43bd-9ee3-095e8e2379f0-kube-api-access-nmwrm\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.216431 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.216450 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metrics-certs\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.216626 4669 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.216684 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metrics-certs podName:ec1f1180-0070-43bd-9ee3-095e8e2379f0 nodeName:}" failed. No retries permitted until 2025-10-10 09:23:55.716664501 +0000 UTC m=+778.732683243 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metrics-certs") pod "speaker-78hzw" (UID: "ec1f1180-0070-43bd-9ee3-095e8e2379f0") : secret "speaker-certs-secret" not found Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.216935 4669 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.216977 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-metrics-certs podName:0eeead67-0b6d-48c7-8756-50e1c1eb5f0b nodeName:}" failed. No retries permitted until 2025-10-10 09:23:55.71696454 +0000 UTC m=+778.732983282 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-metrics-certs") pod "controller-68d546b9d8-xh4cl" (UID: "0eeead67-0b6d-48c7-8756-50e1c1eb5f0b") : secret "controller-certs-secret" not found Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.217855 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metallb-excludel2\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.218384 4669 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.218432 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist podName:ec1f1180-0070-43bd-9ee3-095e8e2379f0 nodeName:}" failed. No retries permitted until 2025-10-10 09:23:55.718422277 +0000 UTC m=+778.734441019 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist") pod "speaker-78hzw" (UID: "ec1f1180-0070-43bd-9ee3-095e8e2379f0") : secret "metallb-memberlist" not found Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.230280 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-cert\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.240453 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmwrm\" (UniqueName: \"kubernetes.io/projected/ec1f1180-0070-43bd-9ee3-095e8e2379f0-kube-api-access-nmwrm\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.241881 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4slzs\" (UniqueName: \"kubernetes.io/projected/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-kube-api-access-4slzs\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.247193 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="b1be725249e1775acac8f350a1439ffc4d89a22bc933af6755f8ceefcd7eb018" exitCode=0 Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.247357 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"b1be725249e1775acac8f350a1439ffc4d89a22bc933af6755f8ceefcd7eb018"} Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.247556 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"460c2cb0159213ee8a9812e151d381634c493f7eb4c849db7277b1ec7a3b7419"} Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.247655 4669 scope.go:117] "RemoveContainer" containerID="86ae923b6b0c46f8077ccd2f9c57221fedfc20fa3778f2e007aa9fa0736e2d76" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.722458 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metrics-certs\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.722798 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.722891 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-metrics-certs\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.723942 4669 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 10 09:23:55 crc kubenswrapper[4669]: E1010 09:23:55.724002 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist podName:ec1f1180-0070-43bd-9ee3-095e8e2379f0 nodeName:}" failed. No retries permitted until 2025-10-10 09:23:56.723987235 +0000 UTC m=+779.740005977 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist") pod "speaker-78hzw" (UID: "ec1f1180-0070-43bd-9ee3-095e8e2379f0") : secret "metallb-memberlist" not found Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.728418 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-metrics-certs\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.731358 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0eeead67-0b6d-48c7-8756-50e1c1eb5f0b-metrics-certs\") pod \"controller-68d546b9d8-xh4cl\" (UID: \"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b\") " pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.795346 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll"] Oct 10 09:23:55 crc kubenswrapper[4669]: I1010 09:23:55.951074 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:56 crc kubenswrapper[4669]: I1010 09:23:56.254674 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerStarted","Data":"9d50d4ffd39b4ed0c538fcd5df33511b26f5f06848a517ce1cda82e20c2f5bb0"} Oct 10 09:23:56 crc kubenswrapper[4669]: I1010 09:23:56.257381 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" event={"ID":"8bd73190-2f9a-45df-8e2d-8579238bf4a3","Type":"ContainerStarted","Data":"894b1d34db42ab6e73e940a47aaaed3e6de3c4765f151611a244eaddeceb9895"} Oct 10 09:23:56 crc kubenswrapper[4669]: I1010 09:23:56.345330 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-xh4cl"] Oct 10 09:23:56 crc kubenswrapper[4669]: W1010 09:23:56.353011 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0eeead67_0b6d_48c7_8756_50e1c1eb5f0b.slice/crio-fdaa0f6f0af07abb685d9a22957e76a00c5bddb6c380c36f866784f499e885d2 WatchSource:0}: Error finding container fdaa0f6f0af07abb685d9a22957e76a00c5bddb6c380c36f866784f499e885d2: Status 404 returned error can't find the container with id fdaa0f6f0af07abb685d9a22957e76a00c5bddb6c380c36f866784f499e885d2 Oct 10 09:23:56 crc kubenswrapper[4669]: I1010 09:23:56.740038 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:56 crc kubenswrapper[4669]: I1010 09:23:56.744696 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec1f1180-0070-43bd-9ee3-095e8e2379f0-memberlist\") pod \"speaker-78hzw\" (UID: \"ec1f1180-0070-43bd-9ee3-095e8e2379f0\") " pod="metallb-system/speaker-78hzw" Oct 10 09:23:56 crc kubenswrapper[4669]: I1010 09:23:56.840705 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-78hzw" Oct 10 09:23:56 crc kubenswrapper[4669]: W1010 09:23:56.886871 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec1f1180_0070_43bd_9ee3_095e8e2379f0.slice/crio-644bf8ad4b93a1a7de630e56c6acf30ea5f484a88f203f534576aff2c6a4e545 WatchSource:0}: Error finding container 644bf8ad4b93a1a7de630e56c6acf30ea5f484a88f203f534576aff2c6a4e545: Status 404 returned error can't find the container with id 644bf8ad4b93a1a7de630e56c6acf30ea5f484a88f203f534576aff2c6a4e545 Oct 10 09:23:57 crc kubenswrapper[4669]: I1010 09:23:57.267975 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-78hzw" event={"ID":"ec1f1180-0070-43bd-9ee3-095e8e2379f0","Type":"ContainerStarted","Data":"3dee3f1969b23a30d8de75135bfd2d93b454f422d12b099f09b8d02a7bf65df1"} Oct 10 09:23:57 crc kubenswrapper[4669]: I1010 09:23:57.268308 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-78hzw" event={"ID":"ec1f1180-0070-43bd-9ee3-095e8e2379f0","Type":"ContainerStarted","Data":"644bf8ad4b93a1a7de630e56c6acf30ea5f484a88f203f534576aff2c6a4e545"} Oct 10 09:23:57 crc kubenswrapper[4669]: I1010 09:23:57.269642 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-xh4cl" event={"ID":"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b","Type":"ContainerStarted","Data":"05dd5a5d6e79c05849c5fc3b78f95e8a568aab6364ba0f959f939d2d64d6c3e8"} Oct 10 09:23:57 crc kubenswrapper[4669]: I1010 09:23:57.269667 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-xh4cl" event={"ID":"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b","Type":"ContainerStarted","Data":"00f79143ad827b5cc6e5fbd4f790b913b2cd04c88c1a15da7e792c6194e3bd27"} Oct 10 09:23:57 crc kubenswrapper[4669]: I1010 09:23:57.269676 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-xh4cl" event={"ID":"0eeead67-0b6d-48c7-8756-50e1c1eb5f0b","Type":"ContainerStarted","Data":"fdaa0f6f0af07abb685d9a22957e76a00c5bddb6c380c36f866784f499e885d2"} Oct 10 09:23:57 crc kubenswrapper[4669]: I1010 09:23:57.270561 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:23:57 crc kubenswrapper[4669]: I1010 09:23:57.291989 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-xh4cl" podStartSLOduration=2.291965067 podStartE2EDuration="2.291965067s" podCreationTimestamp="2025-10-10 09:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:23:57.283775712 +0000 UTC m=+780.299794484" watchObservedRunningTime="2025-10-10 09:23:57.291965067 +0000 UTC m=+780.307983809" Oct 10 09:23:58 crc kubenswrapper[4669]: I1010 09:23:58.277997 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-78hzw" event={"ID":"ec1f1180-0070-43bd-9ee3-095e8e2379f0","Type":"ContainerStarted","Data":"602a5547833140e9dc09f06acbdd7cd5683d54f5d3eabfed79ba74325cae95a5"} Oct 10 09:23:58 crc kubenswrapper[4669]: I1010 09:23:58.278660 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-78hzw" Oct 10 09:24:04 crc kubenswrapper[4669]: I1010 09:24:04.331217 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" event={"ID":"8bd73190-2f9a-45df-8e2d-8579238bf4a3","Type":"ContainerStarted","Data":"ae7e645992952d4eb26e5a22b7ca973e97440556d2f9f714ba4882788213bdc7"} Oct 10 09:24:04 crc kubenswrapper[4669]: I1010 09:24:04.332531 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:24:04 crc kubenswrapper[4669]: I1010 09:24:04.333615 4669 generic.go:334] "Generic (PLEG): container finished" podID="3f8d268c-022e-42b5-8e08-dbdefd895e23" containerID="c3203945b5aa6b5ab4f57a2c7ebfcf8925dc6f7c85f18209e3c33aec116a9c7c" exitCode=0 Oct 10 09:24:04 crc kubenswrapper[4669]: I1010 09:24:04.333641 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerDied","Data":"c3203945b5aa6b5ab4f57a2c7ebfcf8925dc6f7c85f18209e3c33aec116a9c7c"} Oct 10 09:24:04 crc kubenswrapper[4669]: I1010 09:24:04.346491 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-78hzw" podStartSLOduration=9.346473254 podStartE2EDuration="9.346473254s" podCreationTimestamp="2025-10-10 09:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:23:58.295614498 +0000 UTC m=+781.311633240" watchObservedRunningTime="2025-10-10 09:24:04.346473254 +0000 UTC m=+787.362491996" Oct 10 09:24:04 crc kubenswrapper[4669]: I1010 09:24:04.350519 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" podStartSLOduration=2.458205601 podStartE2EDuration="10.350502144s" podCreationTimestamp="2025-10-10 09:23:54 +0000 UTC" firstStartedPulling="2025-10-10 09:23:55.812511463 +0000 UTC m=+778.828530205" lastFinishedPulling="2025-10-10 09:24:03.704808006 +0000 UTC m=+786.720826748" observedRunningTime="2025-10-10 09:24:04.347675602 +0000 UTC m=+787.363694354" watchObservedRunningTime="2025-10-10 09:24:04.350502144 +0000 UTC m=+787.366520876" Oct 10 09:24:05 crc kubenswrapper[4669]: I1010 09:24:05.340382 4669 generic.go:334] "Generic (PLEG): container finished" podID="3f8d268c-022e-42b5-8e08-dbdefd895e23" containerID="b42b674d4279831fb730dde5042588f45272d27fc8403d786dbb202a43a059d5" exitCode=0 Oct 10 09:24:05 crc kubenswrapper[4669]: I1010 09:24:05.340439 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerDied","Data":"b42b674d4279831fb730dde5042588f45272d27fc8403d786dbb202a43a059d5"} Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.354298 4669 generic.go:334] "Generic (PLEG): container finished" podID="3f8d268c-022e-42b5-8e08-dbdefd895e23" containerID="205823c820caf1947ecc333716f40f0512254c92c54b68b9baa93c9d6b4b849e" exitCode=0 Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.354453 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerDied","Data":"205823c820caf1947ecc333716f40f0512254c92c54b68b9baa93c9d6b4b849e"} Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.800317 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tn5hz"] Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.801394 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.822410 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tn5hz"] Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.882135 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-catalog-content\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.882340 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht4fl\" (UniqueName: \"kubernetes.io/projected/9ffd379e-1ed2-42e0-9f04-e37520f5f242-kube-api-access-ht4fl\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.882413 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-utilities\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.983205 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht4fl\" (UniqueName: \"kubernetes.io/projected/9ffd379e-1ed2-42e0-9f04-e37520f5f242-kube-api-access-ht4fl\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.983256 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-utilities\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.983301 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-catalog-content\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.983693 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-catalog-content\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:06 crc kubenswrapper[4669]: I1010 09:24:06.983793 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-utilities\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.007175 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht4fl\" (UniqueName: \"kubernetes.io/projected/9ffd379e-1ed2-42e0-9f04-e37520f5f242-kube-api-access-ht4fl\") pod \"community-operators-tn5hz\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.151686 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.435288 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerStarted","Data":"5646487bb892eedde4ac428accd335af19b23c370d8663a21d120cbaad30d6ef"} Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.435627 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerStarted","Data":"303f709f26dba75db82cce1f61e6cf0db91749cb6c411eb91de70c0c6972c7f7"} Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.435645 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerStarted","Data":"89a469dc6df5c22f6ad5f9a6a05626ce7e280415cd4a1b9858ff033012bc47a2"} Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.435656 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerStarted","Data":"503064576a69f02b31fd2cd243796827c7fc4cf0fba13cb8cbdbe490f6dcbfae"} Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.435665 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerStarted","Data":"70a8ee782fb755d5688b0d3a84e5ea0bd843e6b62b795b1c79db201416b60feb"} Oct 10 09:24:07 crc kubenswrapper[4669]: I1010 09:24:07.649205 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tn5hz"] Oct 10 09:24:08 crc kubenswrapper[4669]: I1010 09:24:08.443699 4669 generic.go:334] "Generic (PLEG): container finished" podID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerID="0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f" exitCode=0 Oct 10 09:24:08 crc kubenswrapper[4669]: I1010 09:24:08.443939 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn5hz" event={"ID":"9ffd379e-1ed2-42e0-9f04-e37520f5f242","Type":"ContainerDied","Data":"0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f"} Oct 10 09:24:08 crc kubenswrapper[4669]: I1010 09:24:08.443973 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn5hz" event={"ID":"9ffd379e-1ed2-42e0-9f04-e37520f5f242","Type":"ContainerStarted","Data":"434a7a99d385dbc6018b51122fe39a4339d9dc620eaf06a84cd7b0728fb887a8"} Oct 10 09:24:08 crc kubenswrapper[4669]: I1010 09:24:08.453208 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-rp4mn" event={"ID":"3f8d268c-022e-42b5-8e08-dbdefd895e23","Type":"ContainerStarted","Data":"bde08a7ecbf938524cdcdb022d7d9ab9172fd8573e5cda71ad01638a56858b6c"} Oct 10 09:24:08 crc kubenswrapper[4669]: I1010 09:24:08.453366 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:24:08 crc kubenswrapper[4669]: I1010 09:24:08.481923 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-rp4mn" podStartSLOduration=6.09123081 podStartE2EDuration="14.481851184s" podCreationTimestamp="2025-10-10 09:23:54 +0000 UTC" firstStartedPulling="2025-10-10 09:23:55.333149346 +0000 UTC m=+778.349168088" lastFinishedPulling="2025-10-10 09:24:03.72376972 +0000 UTC m=+786.739788462" observedRunningTime="2025-10-10 09:24:08.479168337 +0000 UTC m=+791.495187079" watchObservedRunningTime="2025-10-10 09:24:08.481851184 +0000 UTC m=+791.497869936" Oct 10 09:24:09 crc kubenswrapper[4669]: I1010 09:24:09.461744 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn5hz" event={"ID":"9ffd379e-1ed2-42e0-9f04-e37520f5f242","Type":"ContainerStarted","Data":"6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908"} Oct 10 09:24:10 crc kubenswrapper[4669]: I1010 09:24:10.206627 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:24:10 crc kubenswrapper[4669]: I1010 09:24:10.244176 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:24:10 crc kubenswrapper[4669]: I1010 09:24:10.470007 4669 generic.go:334] "Generic (PLEG): container finished" podID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerID="6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908" exitCode=0 Oct 10 09:24:10 crc kubenswrapper[4669]: I1010 09:24:10.470735 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn5hz" event={"ID":"9ffd379e-1ed2-42e0-9f04-e37520f5f242","Type":"ContainerDied","Data":"6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908"} Oct 10 09:24:11 crc kubenswrapper[4669]: I1010 09:24:11.477506 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn5hz" event={"ID":"9ffd379e-1ed2-42e0-9f04-e37520f5f242","Type":"ContainerStarted","Data":"0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05"} Oct 10 09:24:11 crc kubenswrapper[4669]: I1010 09:24:11.497847 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tn5hz" podStartSLOduration=2.870314155 podStartE2EDuration="5.49783202s" podCreationTimestamp="2025-10-10 09:24:06 +0000 UTC" firstStartedPulling="2025-10-10 09:24:08.446109735 +0000 UTC m=+791.462128477" lastFinishedPulling="2025-10-10 09:24:11.0736276 +0000 UTC m=+794.089646342" observedRunningTime="2025-10-10 09:24:11.49351747 +0000 UTC m=+794.509536222" watchObservedRunningTime="2025-10-10 09:24:11.49783202 +0000 UTC m=+794.513850762" Oct 10 09:24:15 crc kubenswrapper[4669]: I1010 09:24:15.199502 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-rgcll" Oct 10 09:24:15 crc kubenswrapper[4669]: I1010 09:24:15.956197 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-xh4cl" Oct 10 09:24:16 crc kubenswrapper[4669]: I1010 09:24:16.844773 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-78hzw" Oct 10 09:24:17 crc kubenswrapper[4669]: I1010 09:24:17.152193 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:17 crc kubenswrapper[4669]: I1010 09:24:17.152391 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:17 crc kubenswrapper[4669]: I1010 09:24:17.190793 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:17 crc kubenswrapper[4669]: I1010 09:24:17.549714 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:17 crc kubenswrapper[4669]: I1010 09:24:17.586157 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tn5hz"] Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.524882 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tn5hz" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="registry-server" containerID="cri-o://0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05" gracePeriod=2 Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.706701 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mch4g"] Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.707568 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mch4g" Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.709892 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-k2jjt" Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.710201 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.710205 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.731540 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mch4g"] Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.873505 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr942\" (UniqueName: \"kubernetes.io/projected/0cf1a943-879b-4c52-a513-11036589f1fc-kube-api-access-zr942\") pod \"openstack-operator-index-mch4g\" (UID: \"0cf1a943-879b-4c52-a513-11036589f1fc\") " pod="openstack-operators/openstack-operator-index-mch4g" Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.975605 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr942\" (UniqueName: \"kubernetes.io/projected/0cf1a943-879b-4c52-a513-11036589f1fc-kube-api-access-zr942\") pod \"openstack-operator-index-mch4g\" (UID: \"0cf1a943-879b-4c52-a513-11036589f1fc\") " pod="openstack-operators/openstack-operator-index-mch4g" Oct 10 09:24:19 crc kubenswrapper[4669]: I1010 09:24:19.996314 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr942\" (UniqueName: \"kubernetes.io/projected/0cf1a943-879b-4c52-a513-11036589f1fc-kube-api-access-zr942\") pod \"openstack-operator-index-mch4g\" (UID: \"0cf1a943-879b-4c52-a513-11036589f1fc\") " pod="openstack-operators/openstack-operator-index-mch4g" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.023681 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mch4g" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.026914 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.177905 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-catalog-content\") pod \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.178175 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-utilities\") pod \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.178214 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ht4fl\" (UniqueName: \"kubernetes.io/projected/9ffd379e-1ed2-42e0-9f04-e37520f5f242-kube-api-access-ht4fl\") pod \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\" (UID: \"9ffd379e-1ed2-42e0-9f04-e37520f5f242\") " Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.180194 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-utilities" (OuterVolumeSpecName: "utilities") pod "9ffd379e-1ed2-42e0-9f04-e37520f5f242" (UID: "9ffd379e-1ed2-42e0-9f04-e37520f5f242"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.182853 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ffd379e-1ed2-42e0-9f04-e37520f5f242-kube-api-access-ht4fl" (OuterVolumeSpecName: "kube-api-access-ht4fl") pod "9ffd379e-1ed2-42e0-9f04-e37520f5f242" (UID: "9ffd379e-1ed2-42e0-9f04-e37520f5f242"). InnerVolumeSpecName "kube-api-access-ht4fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.245525 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ffd379e-1ed2-42e0-9f04-e37520f5f242" (UID: "9ffd379e-1ed2-42e0-9f04-e37520f5f242"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.281461 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.281832 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ffd379e-1ed2-42e0-9f04-e37520f5f242-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.281848 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ht4fl\" (UniqueName: \"kubernetes.io/projected/9ffd379e-1ed2-42e0-9f04-e37520f5f242-kube-api-access-ht4fl\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.505981 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mch4g"] Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.533988 4669 generic.go:334] "Generic (PLEG): container finished" podID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerID="0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05" exitCode=0 Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.534041 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn5hz" event={"ID":"9ffd379e-1ed2-42e0-9f04-e37520f5f242","Type":"ContainerDied","Data":"0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05"} Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.534088 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn5hz" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.534118 4669 scope.go:117] "RemoveContainer" containerID="0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.534102 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn5hz" event={"ID":"9ffd379e-1ed2-42e0-9f04-e37520f5f242","Type":"ContainerDied","Data":"434a7a99d385dbc6018b51122fe39a4339d9dc620eaf06a84cd7b0728fb887a8"} Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.535056 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mch4g" event={"ID":"0cf1a943-879b-4c52-a513-11036589f1fc","Type":"ContainerStarted","Data":"8a8ffee121c51d85fb5e26602339bea1a61547b36aa063c101f5a3f150504e4f"} Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.555276 4669 scope.go:117] "RemoveContainer" containerID="6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.566355 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tn5hz"] Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.570254 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tn5hz"] Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.582545 4669 scope.go:117] "RemoveContainer" containerID="0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.606120 4669 scope.go:117] "RemoveContainer" containerID="0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05" Oct 10 09:24:20 crc kubenswrapper[4669]: E1010 09:24:20.606818 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05\": container with ID starting with 0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05 not found: ID does not exist" containerID="0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.606845 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05"} err="failed to get container status \"0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05\": rpc error: code = NotFound desc = could not find container \"0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05\": container with ID starting with 0208ef2b5879ae833d913a577ddc2733bf0813d14d4c7d3714c4781e2f4abc05 not found: ID does not exist" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.606863 4669 scope.go:117] "RemoveContainer" containerID="6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908" Oct 10 09:24:20 crc kubenswrapper[4669]: E1010 09:24:20.607437 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908\": container with ID starting with 6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908 not found: ID does not exist" containerID="6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.607464 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908"} err="failed to get container status \"6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908\": rpc error: code = NotFound desc = could not find container \"6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908\": container with ID starting with 6cc779e42b42485cead9b84c1a6a38eff67f9f1defc5372111544fece72a6908 not found: ID does not exist" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.607477 4669 scope.go:117] "RemoveContainer" containerID="0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f" Oct 10 09:24:20 crc kubenswrapper[4669]: E1010 09:24:20.608117 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f\": container with ID starting with 0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f not found: ID does not exist" containerID="0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f" Oct 10 09:24:20 crc kubenswrapper[4669]: I1010 09:24:20.608180 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f"} err="failed to get container status \"0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f\": rpc error: code = NotFound desc = could not find container \"0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f\": container with ID starting with 0b89b9d456153b4945160aad5aff92b3d40da1ecc314e975a2a751aff52d847f not found: ID does not exist" Oct 10 09:24:21 crc kubenswrapper[4669]: I1010 09:24:21.803617 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" path="/var/lib/kubelet/pods/9ffd379e-1ed2-42e0-9f04-e37520f5f242/volumes" Oct 10 09:24:23 crc kubenswrapper[4669]: I1010 09:24:23.555276 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mch4g" event={"ID":"0cf1a943-879b-4c52-a513-11036589f1fc","Type":"ContainerStarted","Data":"b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce"} Oct 10 09:24:23 crc kubenswrapper[4669]: I1010 09:24:23.579715 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mch4g" podStartSLOduration=2.428937506 podStartE2EDuration="4.579695979s" podCreationTimestamp="2025-10-10 09:24:19 +0000 UTC" firstStartedPulling="2025-10-10 09:24:20.516091878 +0000 UTC m=+803.532110660" lastFinishedPulling="2025-10-10 09:24:22.666850381 +0000 UTC m=+805.682869133" observedRunningTime="2025-10-10 09:24:23.575443931 +0000 UTC m=+806.591462693" watchObservedRunningTime="2025-10-10 09:24:23.579695979 +0000 UTC m=+806.595714721" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.024640 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mch4g"] Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.208837 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-rp4mn" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.567347 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-mch4g" podUID="0cf1a943-879b-4c52-a513-11036589f1fc" containerName="registry-server" containerID="cri-o://b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce" gracePeriod=2 Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.840798 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2tjnm"] Oct 10 09:24:25 crc kubenswrapper[4669]: E1010 09:24:25.841105 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="extract-content" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.841119 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="extract-content" Oct 10 09:24:25 crc kubenswrapper[4669]: E1010 09:24:25.841137 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="extract-utilities" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.841143 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="extract-utilities" Oct 10 09:24:25 crc kubenswrapper[4669]: E1010 09:24:25.841152 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="registry-server" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.841158 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="registry-server" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.841290 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ffd379e-1ed2-42e0-9f04-e37520f5f242" containerName="registry-server" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.841761 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.843629 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2tjnm"] Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.956789 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6ch5\" (UniqueName: \"kubernetes.io/projected/0e034df9-3c3c-4b15-8d82-2a4179090ed7-kube-api-access-r6ch5\") pod \"openstack-operator-index-2tjnm\" (UID: \"0e034df9-3c3c-4b15-8d82-2a4179090ed7\") " pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:25 crc kubenswrapper[4669]: I1010 09:24:25.965694 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mch4g" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.058273 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr942\" (UniqueName: \"kubernetes.io/projected/0cf1a943-879b-4c52-a513-11036589f1fc-kube-api-access-zr942\") pod \"0cf1a943-879b-4c52-a513-11036589f1fc\" (UID: \"0cf1a943-879b-4c52-a513-11036589f1fc\") " Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.059652 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6ch5\" (UniqueName: \"kubernetes.io/projected/0e034df9-3c3c-4b15-8d82-2a4179090ed7-kube-api-access-r6ch5\") pod \"openstack-operator-index-2tjnm\" (UID: \"0e034df9-3c3c-4b15-8d82-2a4179090ed7\") " pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.067811 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf1a943-879b-4c52-a513-11036589f1fc-kube-api-access-zr942" (OuterVolumeSpecName: "kube-api-access-zr942") pod "0cf1a943-879b-4c52-a513-11036589f1fc" (UID: "0cf1a943-879b-4c52-a513-11036589f1fc"). InnerVolumeSpecName "kube-api-access-zr942". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.076193 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6ch5\" (UniqueName: \"kubernetes.io/projected/0e034df9-3c3c-4b15-8d82-2a4179090ed7-kube-api-access-r6ch5\") pod \"openstack-operator-index-2tjnm\" (UID: \"0e034df9-3c3c-4b15-8d82-2a4179090ed7\") " pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.160654 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.160894 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr942\" (UniqueName: \"kubernetes.io/projected/0cf1a943-879b-4c52-a513-11036589f1fc-kube-api-access-zr942\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.558559 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2tjnm"] Oct 10 09:24:26 crc kubenswrapper[4669]: W1010 09:24:26.560518 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e034df9_3c3c_4b15_8d82_2a4179090ed7.slice/crio-9d93bdb162b04902912bed001ba1d56a92bca9457e0914a02db27b16b8113f7a WatchSource:0}: Error finding container 9d93bdb162b04902912bed001ba1d56a92bca9457e0914a02db27b16b8113f7a: Status 404 returned error can't find the container with id 9d93bdb162b04902912bed001ba1d56a92bca9457e0914a02db27b16b8113f7a Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.573818 4669 generic.go:334] "Generic (PLEG): container finished" podID="0cf1a943-879b-4c52-a513-11036589f1fc" containerID="b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce" exitCode=0 Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.573913 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mch4g" event={"ID":"0cf1a943-879b-4c52-a513-11036589f1fc","Type":"ContainerDied","Data":"b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce"} Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.573977 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mch4g" event={"ID":"0cf1a943-879b-4c52-a513-11036589f1fc","Type":"ContainerDied","Data":"8a8ffee121c51d85fb5e26602339bea1a61547b36aa063c101f5a3f150504e4f"} Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.573998 4669 scope.go:117] "RemoveContainer" containerID="b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.574164 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mch4g" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.575974 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2tjnm" event={"ID":"0e034df9-3c3c-4b15-8d82-2a4179090ed7","Type":"ContainerStarted","Data":"9d93bdb162b04902912bed001ba1d56a92bca9457e0914a02db27b16b8113f7a"} Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.594930 4669 scope.go:117] "RemoveContainer" containerID="b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce" Oct 10 09:24:26 crc kubenswrapper[4669]: E1010 09:24:26.595353 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce\": container with ID starting with b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce not found: ID does not exist" containerID="b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.595400 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce"} err="failed to get container status \"b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce\": rpc error: code = NotFound desc = could not find container \"b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce\": container with ID starting with b1045a48120f4148f551dfa769b9b1ecbfa2fa21f8081eae8c0f2b47bb0640ce not found: ID does not exist" Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.608336 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mch4g"] Oct 10 09:24:26 crc kubenswrapper[4669]: I1010 09:24:26.612104 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-mch4g"] Oct 10 09:24:27 crc kubenswrapper[4669]: I1010 09:24:27.584526 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2tjnm" event={"ID":"0e034df9-3c3c-4b15-8d82-2a4179090ed7","Type":"ContainerStarted","Data":"9a2e2a25b5c4be2ae691f833c858ee8e960a8b50ea664ae18ae85da4161b0254"} Oct 10 09:24:27 crc kubenswrapper[4669]: I1010 09:24:27.802742 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf1a943-879b-4c52-a513-11036589f1fc" path="/var/lib/kubelet/pods/0cf1a943-879b-4c52-a513-11036589f1fc/volumes" Oct 10 09:24:36 crc kubenswrapper[4669]: I1010 09:24:36.161262 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:36 crc kubenswrapper[4669]: I1010 09:24:36.162354 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:36 crc kubenswrapper[4669]: I1010 09:24:36.193613 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:36 crc kubenswrapper[4669]: I1010 09:24:36.208106 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2tjnm" podStartSLOduration=10.679729043 podStartE2EDuration="11.208088269s" podCreationTimestamp="2025-10-10 09:24:25 +0000 UTC" firstStartedPulling="2025-10-10 09:24:26.565113934 +0000 UTC m=+809.581132676" lastFinishedPulling="2025-10-10 09:24:27.09347316 +0000 UTC m=+810.109491902" observedRunningTime="2025-10-10 09:24:27.600656289 +0000 UTC m=+810.616675051" watchObservedRunningTime="2025-10-10 09:24:36.208088269 +0000 UTC m=+819.224107011" Oct 10 09:24:36 crc kubenswrapper[4669]: I1010 09:24:36.671757 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-2tjnm" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.262216 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn"] Oct 10 09:24:37 crc kubenswrapper[4669]: E1010 09:24:37.262470 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1a943-879b-4c52-a513-11036589f1fc" containerName="registry-server" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.262483 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1a943-879b-4c52-a513-11036589f1fc" containerName="registry-server" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.262641 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf1a943-879b-4c52-a513-11036589f1fc" containerName="registry-server" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.263603 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.266308 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-sgtrp" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.276986 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn"] Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.402626 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xdfj\" (UniqueName: \"kubernetes.io/projected/236d5734-1ffa-4791-abf6-2fe59ade1eeb-kube-api-access-7xdfj\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.402702 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-util\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.402828 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-bundle\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.504606 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xdfj\" (UniqueName: \"kubernetes.io/projected/236d5734-1ffa-4791-abf6-2fe59ade1eeb-kube-api-access-7xdfj\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.504673 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-util\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.504736 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-bundle\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.505191 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-bundle\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.505655 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-util\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.531994 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xdfj\" (UniqueName: \"kubernetes.io/projected/236d5734-1ffa-4791-abf6-2fe59ade1eeb-kube-api-access-7xdfj\") pod \"f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.583966 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:37 crc kubenswrapper[4669]: I1010 09:24:37.973572 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn"] Oct 10 09:24:38 crc kubenswrapper[4669]: I1010 09:24:38.663938 4669 generic.go:334] "Generic (PLEG): container finished" podID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerID="7d112f7d01a32dd19f04b41ef0b4995ababf9da74b8e16ab53807b834d39cae3" exitCode=0 Oct 10 09:24:38 crc kubenswrapper[4669]: I1010 09:24:38.664013 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" event={"ID":"236d5734-1ffa-4791-abf6-2fe59ade1eeb","Type":"ContainerDied","Data":"7d112f7d01a32dd19f04b41ef0b4995ababf9da74b8e16ab53807b834d39cae3"} Oct 10 09:24:38 crc kubenswrapper[4669]: I1010 09:24:38.664435 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" event={"ID":"236d5734-1ffa-4791-abf6-2fe59ade1eeb","Type":"ContainerStarted","Data":"408444aaeb372c4007fae57aecd24d0fa13743a553bf7c1b4fb597687568133d"} Oct 10 09:24:39 crc kubenswrapper[4669]: I1010 09:24:39.672241 4669 generic.go:334] "Generic (PLEG): container finished" podID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerID="f5cfbb0ae82136631b62c427f1a058eff0b05dbd2a1fb247cadc40016ef85406" exitCode=0 Oct 10 09:24:39 crc kubenswrapper[4669]: I1010 09:24:39.672290 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" event={"ID":"236d5734-1ffa-4791-abf6-2fe59ade1eeb","Type":"ContainerDied","Data":"f5cfbb0ae82136631b62c427f1a058eff0b05dbd2a1fb247cadc40016ef85406"} Oct 10 09:24:40 crc kubenswrapper[4669]: I1010 09:24:40.682870 4669 generic.go:334] "Generic (PLEG): container finished" podID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerID="19dcc39f54d74e3a79030c0e91ba66817048f2f4671ac00ba294317ea197721f" exitCode=0 Oct 10 09:24:40 crc kubenswrapper[4669]: I1010 09:24:40.683016 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" event={"ID":"236d5734-1ffa-4791-abf6-2fe59ade1eeb","Type":"ContainerDied","Data":"19dcc39f54d74e3a79030c0e91ba66817048f2f4671ac00ba294317ea197721f"} Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.004964 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.172370 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xdfj\" (UniqueName: \"kubernetes.io/projected/236d5734-1ffa-4791-abf6-2fe59ade1eeb-kube-api-access-7xdfj\") pod \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.172440 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-util\") pod \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.172502 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-bundle\") pod \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\" (UID: \"236d5734-1ffa-4791-abf6-2fe59ade1eeb\") " Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.173356 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-bundle" (OuterVolumeSpecName: "bundle") pod "236d5734-1ffa-4791-abf6-2fe59ade1eeb" (UID: "236d5734-1ffa-4791-abf6-2fe59ade1eeb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.181334 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/236d5734-1ffa-4791-abf6-2fe59ade1eeb-kube-api-access-7xdfj" (OuterVolumeSpecName: "kube-api-access-7xdfj") pod "236d5734-1ffa-4791-abf6-2fe59ade1eeb" (UID: "236d5734-1ffa-4791-abf6-2fe59ade1eeb"). InnerVolumeSpecName "kube-api-access-7xdfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.187525 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-util" (OuterVolumeSpecName: "util") pod "236d5734-1ffa-4791-abf6-2fe59ade1eeb" (UID: "236d5734-1ffa-4791-abf6-2fe59ade1eeb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.274526 4669 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-util\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.274551 4669 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/236d5734-1ffa-4791-abf6-2fe59ade1eeb-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.274562 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xdfj\" (UniqueName: \"kubernetes.io/projected/236d5734-1ffa-4791-abf6-2fe59ade1eeb-kube-api-access-7xdfj\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.694545 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" event={"ID":"236d5734-1ffa-4791-abf6-2fe59ade1eeb","Type":"ContainerDied","Data":"408444aaeb372c4007fae57aecd24d0fa13743a553bf7c1b4fb597687568133d"} Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.694602 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="408444aaeb372c4007fae57aecd24d0fa13743a553bf7c1b4fb597687568133d" Oct 10 09:24:42 crc kubenswrapper[4669]: I1010 09:24:42.694632 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.043287 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2m78c"] Oct 10 09:24:44 crc kubenswrapper[4669]: E1010 09:24:44.044571 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerName="util" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.044643 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerName="util" Oct 10 09:24:44 crc kubenswrapper[4669]: E1010 09:24:44.044737 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerName="pull" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.044755 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerName="pull" Oct 10 09:24:44 crc kubenswrapper[4669]: E1010 09:24:44.044788 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerName="extract" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.044834 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerName="extract" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.045394 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="236d5734-1ffa-4791-abf6-2fe59ade1eeb" containerName="extract" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.050346 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.073955 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m78c"] Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.196382 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-catalog-content\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.196424 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjgpw\" (UniqueName: \"kubernetes.io/projected/093e5580-e82d-4fbc-b1ea-0947041c9713-kube-api-access-xjgpw\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.196454 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-utilities\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.297735 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-catalog-content\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.297807 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjgpw\" (UniqueName: \"kubernetes.io/projected/093e5580-e82d-4fbc-b1ea-0947041c9713-kube-api-access-xjgpw\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.297867 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-utilities\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.298491 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-utilities\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.298756 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-catalog-content\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.321391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjgpw\" (UniqueName: \"kubernetes.io/projected/093e5580-e82d-4fbc-b1ea-0947041c9713-kube-api-access-xjgpw\") pod \"redhat-marketplace-2m78c\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.388347 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:44 crc kubenswrapper[4669]: I1010 09:24:44.792814 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m78c"] Oct 10 09:24:45 crc kubenswrapper[4669]: I1010 09:24:45.716120 4669 generic.go:334] "Generic (PLEG): container finished" podID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerID="22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752" exitCode=0 Oct 10 09:24:45 crc kubenswrapper[4669]: I1010 09:24:45.716191 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m78c" event={"ID":"093e5580-e82d-4fbc-b1ea-0947041c9713","Type":"ContainerDied","Data":"22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752"} Oct 10 09:24:45 crc kubenswrapper[4669]: I1010 09:24:45.716413 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m78c" event={"ID":"093e5580-e82d-4fbc-b1ea-0947041c9713","Type":"ContainerStarted","Data":"93b6016144a0101b05d0aa6e549f1eed77389c71348da0369decbc7ebe5c78b8"} Oct 10 09:24:46 crc kubenswrapper[4669]: I1010 09:24:46.723874 4669 generic.go:334] "Generic (PLEG): container finished" podID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerID="9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1" exitCode=0 Oct 10 09:24:46 crc kubenswrapper[4669]: I1010 09:24:46.723929 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m78c" event={"ID":"093e5580-e82d-4fbc-b1ea-0947041c9713","Type":"ContainerDied","Data":"9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1"} Oct 10 09:24:47 crc kubenswrapper[4669]: I1010 09:24:47.731951 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m78c" event={"ID":"093e5580-e82d-4fbc-b1ea-0947041c9713","Type":"ContainerStarted","Data":"76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf"} Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.743363 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2m78c" podStartSLOduration=3.190284993 podStartE2EDuration="4.743345532s" podCreationTimestamp="2025-10-10 09:24:44 +0000 UTC" firstStartedPulling="2025-10-10 09:24:45.717698022 +0000 UTC m=+828.733716764" lastFinishedPulling="2025-10-10 09:24:47.270758551 +0000 UTC m=+830.286777303" observedRunningTime="2025-10-10 09:24:47.753051744 +0000 UTC m=+830.769070486" watchObservedRunningTime="2025-10-10 09:24:48.743345532 +0000 UTC m=+831.759364274" Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.746750 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v"] Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.747761 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.752657 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-rbw95" Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.781236 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v"] Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.858282 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j2wx\" (UniqueName: \"kubernetes.io/projected/d6df673a-6169-4bcc-bf13-98eb9e3a20f9-kube-api-access-4j2wx\") pod \"openstack-operator-controller-operator-599bffcb5d-2bj6v\" (UID: \"d6df673a-6169-4bcc-bf13-98eb9e3a20f9\") " pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.960100 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j2wx\" (UniqueName: \"kubernetes.io/projected/d6df673a-6169-4bcc-bf13-98eb9e3a20f9-kube-api-access-4j2wx\") pod \"openstack-operator-controller-operator-599bffcb5d-2bj6v\" (UID: \"d6df673a-6169-4bcc-bf13-98eb9e3a20f9\") " pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" Oct 10 09:24:48 crc kubenswrapper[4669]: I1010 09:24:48.979288 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j2wx\" (UniqueName: \"kubernetes.io/projected/d6df673a-6169-4bcc-bf13-98eb9e3a20f9-kube-api-access-4j2wx\") pod \"openstack-operator-controller-operator-599bffcb5d-2bj6v\" (UID: \"d6df673a-6169-4bcc-bf13-98eb9e3a20f9\") " pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" Oct 10 09:24:49 crc kubenswrapper[4669]: I1010 09:24:49.062013 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" Oct 10 09:24:49 crc kubenswrapper[4669]: I1010 09:24:49.545396 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v"] Oct 10 09:24:49 crc kubenswrapper[4669]: I1010 09:24:49.741770 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" event={"ID":"d6df673a-6169-4bcc-bf13-98eb9e3a20f9","Type":"ContainerStarted","Data":"856bbaa0c8ee24fcf08efd20e2c0f30903250ad9cab0ee773c6f4fa718d19bce"} Oct 10 09:24:54 crc kubenswrapper[4669]: I1010 09:24:54.390120 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:54 crc kubenswrapper[4669]: I1010 09:24:54.391193 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:54 crc kubenswrapper[4669]: I1010 09:24:54.436044 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:54 crc kubenswrapper[4669]: I1010 09:24:54.776169 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" event={"ID":"d6df673a-6169-4bcc-bf13-98eb9e3a20f9","Type":"ContainerStarted","Data":"9d260699ad6d48ca9e737fada18d69ea17a3b076d97786338c49486371d0dd8d"} Oct 10 09:24:54 crc kubenswrapper[4669]: I1010 09:24:54.825655 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:56 crc kubenswrapper[4669]: I1010 09:24:56.803408 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" event={"ID":"d6df673a-6169-4bcc-bf13-98eb9e3a20f9","Type":"ContainerStarted","Data":"bd60a62b35be1ba8735a0e0c3679e1c1fcddb7196961073072b5f2def07dd5da"} Oct 10 09:24:56 crc kubenswrapper[4669]: I1010 09:24:56.804026 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" Oct 10 09:24:56 crc kubenswrapper[4669]: I1010 09:24:56.822369 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m78c"] Oct 10 09:24:56 crc kubenswrapper[4669]: I1010 09:24:56.845980 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" podStartSLOduration=1.8950787340000002 podStartE2EDuration="8.845961322s" podCreationTimestamp="2025-10-10 09:24:48 +0000 UTC" firstStartedPulling="2025-10-10 09:24:49.554513025 +0000 UTC m=+832.570531767" lastFinishedPulling="2025-10-10 09:24:56.505395613 +0000 UTC m=+839.521414355" observedRunningTime="2025-10-10 09:24:56.838156209 +0000 UTC m=+839.854174961" watchObservedRunningTime="2025-10-10 09:24:56.845961322 +0000 UTC m=+839.861980064" Oct 10 09:24:57 crc kubenswrapper[4669]: I1010 09:24:57.810507 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2m78c" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="registry-server" containerID="cri-o://76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf" gracePeriod=2 Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.216147 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.282807 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-utilities\") pod \"093e5580-e82d-4fbc-b1ea-0947041c9713\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.282874 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjgpw\" (UniqueName: \"kubernetes.io/projected/093e5580-e82d-4fbc-b1ea-0947041c9713-kube-api-access-xjgpw\") pod \"093e5580-e82d-4fbc-b1ea-0947041c9713\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.282935 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-catalog-content\") pod \"093e5580-e82d-4fbc-b1ea-0947041c9713\" (UID: \"093e5580-e82d-4fbc-b1ea-0947041c9713\") " Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.294372 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-utilities" (OuterVolumeSpecName: "utilities") pod "093e5580-e82d-4fbc-b1ea-0947041c9713" (UID: "093e5580-e82d-4fbc-b1ea-0947041c9713"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.298847 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093e5580-e82d-4fbc-b1ea-0947041c9713-kube-api-access-xjgpw" (OuterVolumeSpecName: "kube-api-access-xjgpw") pod "093e5580-e82d-4fbc-b1ea-0947041c9713" (UID: "093e5580-e82d-4fbc-b1ea-0947041c9713"). InnerVolumeSpecName "kube-api-access-xjgpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.301745 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "093e5580-e82d-4fbc-b1ea-0947041c9713" (UID: "093e5580-e82d-4fbc-b1ea-0947041c9713"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.394230 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjgpw\" (UniqueName: \"kubernetes.io/projected/093e5580-e82d-4fbc-b1ea-0947041c9713-kube-api-access-xjgpw\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.394260 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.394272 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/093e5580-e82d-4fbc-b1ea-0947041c9713-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.817099 4669 generic.go:334] "Generic (PLEG): container finished" podID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerID="76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf" exitCode=0 Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.817144 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m78c" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.817168 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m78c" event={"ID":"093e5580-e82d-4fbc-b1ea-0947041c9713","Type":"ContainerDied","Data":"76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf"} Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.817250 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m78c" event={"ID":"093e5580-e82d-4fbc-b1ea-0947041c9713","Type":"ContainerDied","Data":"93b6016144a0101b05d0aa6e549f1eed77389c71348da0369decbc7ebe5c78b8"} Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.817275 4669 scope.go:117] "RemoveContainer" containerID="76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.843443 4669 scope.go:117] "RemoveContainer" containerID="9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.844085 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m78c"] Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.848807 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m78c"] Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.861831 4669 scope.go:117] "RemoveContainer" containerID="22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.877620 4669 scope.go:117] "RemoveContainer" containerID="76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf" Oct 10 09:24:58 crc kubenswrapper[4669]: E1010 09:24:58.878068 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf\": container with ID starting with 76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf not found: ID does not exist" containerID="76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.878126 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf"} err="failed to get container status \"76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf\": rpc error: code = NotFound desc = could not find container \"76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf\": container with ID starting with 76ceaf322ea425ca4e9dfb60340f9f8c1497bbbe482d9e048d0487d9cbb751cf not found: ID does not exist" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.878150 4669 scope.go:117] "RemoveContainer" containerID="9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1" Oct 10 09:24:58 crc kubenswrapper[4669]: E1010 09:24:58.878550 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1\": container with ID starting with 9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1 not found: ID does not exist" containerID="9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.878581 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1"} err="failed to get container status \"9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1\": rpc error: code = NotFound desc = could not find container \"9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1\": container with ID starting with 9d906925d9baca6c844fa76d677efea494b4165aa10e14bc06ec63d1ca4f1ce1 not found: ID does not exist" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.878615 4669 scope.go:117] "RemoveContainer" containerID="22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752" Oct 10 09:24:58 crc kubenswrapper[4669]: E1010 09:24:58.878874 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752\": container with ID starting with 22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752 not found: ID does not exist" containerID="22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752" Oct 10 09:24:58 crc kubenswrapper[4669]: I1010 09:24:58.878918 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752"} err="failed to get container status \"22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752\": rpc error: code = NotFound desc = could not find container \"22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752\": container with ID starting with 22caac31e933096c4f431185c4b5617db8d3fbfd6a3bfc30f1ad0c530d513752 not found: ID does not exist" Oct 10 09:24:59 crc kubenswrapper[4669]: I1010 09:24:59.065137 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-599bffcb5d-2bj6v" Oct 10 09:24:59 crc kubenswrapper[4669]: I1010 09:24:59.809756 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" path="/var/lib/kubelet/pods/093e5580-e82d-4fbc-b1ea-0947041c9713/volumes" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.272028 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wq94g"] Oct 10 09:25:05 crc kubenswrapper[4669]: E1010 09:25:05.273626 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="extract-utilities" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.273712 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="extract-utilities" Oct 10 09:25:05 crc kubenswrapper[4669]: E1010 09:25:05.273771 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="extract-content" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.273826 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="extract-content" Oct 10 09:25:05 crc kubenswrapper[4669]: E1010 09:25:05.273879 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="registry-server" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.273933 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="registry-server" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.274089 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="093e5580-e82d-4fbc-b1ea-0947041c9713" containerName="registry-server" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.274955 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.293781 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wq94g"] Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.383845 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp98x\" (UniqueName: \"kubernetes.io/projected/5c07ede9-cc73-43c5-812b-0db964c66349-kube-api-access-pp98x\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.383905 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-utilities\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.383955 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-catalog-content\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.485073 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-catalog-content\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.485142 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp98x\" (UniqueName: \"kubernetes.io/projected/5c07ede9-cc73-43c5-812b-0db964c66349-kube-api-access-pp98x\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.485176 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-utilities\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.485578 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-utilities\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.485798 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-catalog-content\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.522860 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp98x\" (UniqueName: \"kubernetes.io/projected/5c07ede9-cc73-43c5-812b-0db964c66349-kube-api-access-pp98x\") pod \"certified-operators-wq94g\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:05 crc kubenswrapper[4669]: I1010 09:25:05.594563 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:06 crc kubenswrapper[4669]: I1010 09:25:06.160563 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wq94g"] Oct 10 09:25:06 crc kubenswrapper[4669]: I1010 09:25:06.861704 4669 generic.go:334] "Generic (PLEG): container finished" podID="5c07ede9-cc73-43c5-812b-0db964c66349" containerID="f16b6d40f7eb99af350d77b87666dff9944fc10a492ea30dffec028e3fa0cecb" exitCode=0 Oct 10 09:25:06 crc kubenswrapper[4669]: I1010 09:25:06.861905 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq94g" event={"ID":"5c07ede9-cc73-43c5-812b-0db964c66349","Type":"ContainerDied","Data":"f16b6d40f7eb99af350d77b87666dff9944fc10a492ea30dffec028e3fa0cecb"} Oct 10 09:25:06 crc kubenswrapper[4669]: I1010 09:25:06.861943 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq94g" event={"ID":"5c07ede9-cc73-43c5-812b-0db964c66349","Type":"ContainerStarted","Data":"8949dcf0569735706770c8a1fa66cacf3c83356b14d80e3793f6a4eeb272de07"} Oct 10 09:25:07 crc kubenswrapper[4669]: I1010 09:25:07.868242 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq94g" event={"ID":"5c07ede9-cc73-43c5-812b-0db964c66349","Type":"ContainerStarted","Data":"a15a2ee89bb8f9640d1ca743a3e3b21e4f7e7eaabfb7bf83b2bec4bf3eaed075"} Oct 10 09:25:08 crc kubenswrapper[4669]: I1010 09:25:08.875978 4669 generic.go:334] "Generic (PLEG): container finished" podID="5c07ede9-cc73-43c5-812b-0db964c66349" containerID="a15a2ee89bb8f9640d1ca743a3e3b21e4f7e7eaabfb7bf83b2bec4bf3eaed075" exitCode=0 Oct 10 09:25:08 crc kubenswrapper[4669]: I1010 09:25:08.876048 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq94g" event={"ID":"5c07ede9-cc73-43c5-812b-0db964c66349","Type":"ContainerDied","Data":"a15a2ee89bb8f9640d1ca743a3e3b21e4f7e7eaabfb7bf83b2bec4bf3eaed075"} Oct 10 09:25:09 crc kubenswrapper[4669]: I1010 09:25:09.882818 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq94g" event={"ID":"5c07ede9-cc73-43c5-812b-0db964c66349","Type":"ContainerStarted","Data":"9d1fcd2ed9f0e77f9e5243cde468c6399dcd7a6b506df4e2532297d0ecb1b77a"} Oct 10 09:25:09 crc kubenswrapper[4669]: I1010 09:25:09.904509 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wq94g" podStartSLOduration=2.296712982 podStartE2EDuration="4.904490374s" podCreationTimestamp="2025-10-10 09:25:05 +0000 UTC" firstStartedPulling="2025-10-10 09:25:06.863376699 +0000 UTC m=+849.879395441" lastFinishedPulling="2025-10-10 09:25:09.471154091 +0000 UTC m=+852.487172833" observedRunningTime="2025-10-10 09:25:09.901229267 +0000 UTC m=+852.917248009" watchObservedRunningTime="2025-10-10 09:25:09.904490374 +0000 UTC m=+852.920509116" Oct 10 09:25:15 crc kubenswrapper[4669]: I1010 09:25:15.609433 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:15 crc kubenswrapper[4669]: I1010 09:25:15.609793 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:15 crc kubenswrapper[4669]: I1010 09:25:15.649728 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:15 crc kubenswrapper[4669]: I1010 09:25:15.953021 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.005155 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wq94g"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.093575 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.094788 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.096711 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fvvgh" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.110076 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.118091 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.119001 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.121524 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-hb64k" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.123714 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.124808 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.126705 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cmm74" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.159272 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.160236 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.161827 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gd4lh" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.172334 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.175468 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.211759 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.223290 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqw45\" (UniqueName: \"kubernetes.io/projected/1a5e08a1-2d18-4382-8129-be1376c40bde-kube-api-access-nqw45\") pod \"designate-operator-controller-manager-85d5d9dd78-pn8xd\" (UID: \"1a5e08a1-2d18-4382-8129-be1376c40bde\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.223354 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m6xl\" (UniqueName: \"kubernetes.io/projected/7dcc64bf-ac2d-4357-adcb-b6c146462464-kube-api-access-4m6xl\") pod \"cinder-operator-controller-manager-7b7fb68549-jzps8\" (UID: \"7dcc64bf-ac2d-4357-adcb-b6c146462464\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.223406 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4swb\" (UniqueName: \"kubernetes.io/projected/affd00ea-eac5-4c5f-b452-c8bb0e17aba8-kube-api-access-k4swb\") pod \"barbican-operator-controller-manager-658bdf4b74-sbc5v\" (UID: \"affd00ea-eac5-4c5f-b452-c8bb0e17aba8\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.223445 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rss92\" (UniqueName: \"kubernetes.io/projected/55f7e993-3dd8-4acf-9716-47cef7f73ef8-kube-api-access-rss92\") pod \"glance-operator-controller-manager-84b9b84486-pfp4n\" (UID: \"55f7e993-3dd8-4acf-9716-47cef7f73ef8\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.275685 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.277259 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.283932 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-hj2zp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.312253 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-czldp"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.313153 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.313902 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.313971 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.321942 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.322159 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kgmk6" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.322366 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-stkz4" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.342189 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqw45\" (UniqueName: \"kubernetes.io/projected/1a5e08a1-2d18-4382-8129-be1376c40bde-kube-api-access-nqw45\") pod \"designate-operator-controller-manager-85d5d9dd78-pn8xd\" (UID: \"1a5e08a1-2d18-4382-8129-be1376c40bde\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.342272 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m6xl\" (UniqueName: \"kubernetes.io/projected/7dcc64bf-ac2d-4357-adcb-b6c146462464-kube-api-access-4m6xl\") pod \"cinder-operator-controller-manager-7b7fb68549-jzps8\" (UID: \"7dcc64bf-ac2d-4357-adcb-b6c146462464\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.342311 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4swb\" (UniqueName: \"kubernetes.io/projected/affd00ea-eac5-4c5f-b452-c8bb0e17aba8-kube-api-access-k4swb\") pod \"barbican-operator-controller-manager-658bdf4b74-sbc5v\" (UID: \"affd00ea-eac5-4c5f-b452-c8bb0e17aba8\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.342343 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rss92\" (UniqueName: \"kubernetes.io/projected/55f7e993-3dd8-4acf-9716-47cef7f73ef8-kube-api-access-rss92\") pod \"glance-operator-controller-manager-84b9b84486-pfp4n\" (UID: \"55f7e993-3dd8-4acf-9716-47cef7f73ef8\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.344368 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.368681 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.372409 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.379050 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-czldp"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.380216 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-ljlhc" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.389990 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqw45\" (UniqueName: \"kubernetes.io/projected/1a5e08a1-2d18-4382-8129-be1376c40bde-kube-api-access-nqw45\") pod \"designate-operator-controller-manager-85d5d9dd78-pn8xd\" (UID: \"1a5e08a1-2d18-4382-8129-be1376c40bde\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.392718 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.397207 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rss92\" (UniqueName: \"kubernetes.io/projected/55f7e993-3dd8-4acf-9716-47cef7f73ef8-kube-api-access-rss92\") pod \"glance-operator-controller-manager-84b9b84486-pfp4n\" (UID: \"55f7e993-3dd8-4acf-9716-47cef7f73ef8\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.402693 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.405125 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.406213 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.416524 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4swb\" (UniqueName: \"kubernetes.io/projected/affd00ea-eac5-4c5f-b452-c8bb0e17aba8-kube-api-access-k4swb\") pod \"barbican-operator-controller-manager-658bdf4b74-sbc5v\" (UID: \"affd00ea-eac5-4c5f-b452-c8bb0e17aba8\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.419885 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m6xl\" (UniqueName: \"kubernetes.io/projected/7dcc64bf-ac2d-4357-adcb-b6c146462464-kube-api-access-4m6xl\") pod \"cinder-operator-controller-manager-7b7fb68549-jzps8\" (UID: \"7dcc64bf-ac2d-4357-adcb-b6c146462464\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.420297 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-fmn64" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.420641 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.421699 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.423974 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-lwv8f" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.445822 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.446826 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.447557 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.448463 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlgsf\" (UniqueName: \"kubernetes.io/projected/bd3dd74a-3c58-4785-854e-400c7ba726de-kube-api-access-hlgsf\") pod \"heat-operator-controller-manager-858f76bbdd-pdnnc\" (UID: \"bd3dd74a-3c58-4785-854e-400c7ba726de\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.448489 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7528\" (UniqueName: \"kubernetes.io/projected/2132b635-c43e-4c4d-9074-37e81b6345f2-kube-api-access-w7528\") pod \"ironic-operator-controller-manager-9c5c78d49-k5nxp\" (UID: \"2132b635-c43e-4c4d-9074-37e81b6345f2\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.448513 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/270b1e97-6178-44d9-8d9f-ed5e4dc02000-cert\") pod \"infra-operator-controller-manager-656bcbd775-czldp\" (UID: \"270b1e97-6178-44d9-8d9f-ed5e4dc02000\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.448561 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb6z9\" (UniqueName: \"kubernetes.io/projected/909f202d-6571-4743-9314-26cc27e94c77-kube-api-access-tb6z9\") pod \"horizon-operator-controller-manager-7ffbcb7588-96278\" (UID: \"909f202d-6571-4743-9314-26cc27e94c77\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.448612 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5fvj\" (UniqueName: \"kubernetes.io/projected/270b1e97-6178-44d9-8d9f-ed5e4dc02000-kube-api-access-t5fvj\") pod \"infra-operator-controller-manager-656bcbd775-czldp\" (UID: \"270b1e97-6178-44d9-8d9f-ed5e4dc02000\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.453133 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-kdksn" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.453707 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.458937 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.469637 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.470553 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.482936 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bnvzm" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.499146 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.512949 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.523918 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.545815 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.547294 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.551838 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rbc4m" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.557165 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/270b1e97-6178-44d9-8d9f-ed5e4dc02000-cert\") pod \"infra-operator-controller-manager-656bcbd775-czldp\" (UID: \"270b1e97-6178-44d9-8d9f-ed5e4dc02000\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.557528 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzlrf\" (UniqueName: \"kubernetes.io/projected/82353170-7f06-43e4-966b-20d0d234f0f2-kube-api-access-rzlrf\") pod \"mariadb-operator-controller-manager-f9fb45f8f-n72xm\" (UID: \"82353170-7f06-43e4-966b-20d0d234f0f2\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.557643 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb6z9\" (UniqueName: \"kubernetes.io/projected/909f202d-6571-4743-9314-26cc27e94c77-kube-api-access-tb6z9\") pod \"horizon-operator-controller-manager-7ffbcb7588-96278\" (UID: \"909f202d-6571-4743-9314-26cc27e94c77\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.557755 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrdc7\" (UniqueName: \"kubernetes.io/projected/27b6cee3-b0a3-4116-b9bb-299a81bba403-kube-api-access-mrdc7\") pod \"keystone-operator-controller-manager-55b6b7c7b8-6zprx\" (UID: \"27b6cee3-b0a3-4116-b9bb-299a81bba403\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.557858 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kr5z\" (UniqueName: \"kubernetes.io/projected/6941897d-fd2e-4ff3-ad50-aeba82eb86c2-kube-api-access-4kr5z\") pod \"manila-operator-controller-manager-5f67fbc655-q52hz\" (UID: \"6941897d-fd2e-4ff3-ad50-aeba82eb86c2\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.557968 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5fvj\" (UniqueName: \"kubernetes.io/projected/270b1e97-6178-44d9-8d9f-ed5e4dc02000-kube-api-access-t5fvj\") pod \"infra-operator-controller-manager-656bcbd775-czldp\" (UID: \"270b1e97-6178-44d9-8d9f-ed5e4dc02000\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.558067 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw2t2\" (UniqueName: \"kubernetes.io/projected/bdebd762-e639-4e50-820d-cd32885a1341-kube-api-access-rw2t2\") pod \"neutron-operator-controller-manager-79d585cb66-kwn4g\" (UID: \"bdebd762-e639-4e50-820d-cd32885a1341\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.558226 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlgsf\" (UniqueName: \"kubernetes.io/projected/bd3dd74a-3c58-4785-854e-400c7ba726de-kube-api-access-hlgsf\") pod \"heat-operator-controller-manager-858f76bbdd-pdnnc\" (UID: \"bd3dd74a-3c58-4785-854e-400c7ba726de\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.558336 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7528\" (UniqueName: \"kubernetes.io/projected/2132b635-c43e-4c4d-9074-37e81b6345f2-kube-api-access-w7528\") pod \"ironic-operator-controller-manager-9c5c78d49-k5nxp\" (UID: \"2132b635-c43e-4c4d-9074-37e81b6345f2\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" Oct 10 09:25:16 crc kubenswrapper[4669]: E1010 09:25:16.557453 4669 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 10 09:25:16 crc kubenswrapper[4669]: E1010 09:25:16.558576 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/270b1e97-6178-44d9-8d9f-ed5e4dc02000-cert podName:270b1e97-6178-44d9-8d9f-ed5e4dc02000 nodeName:}" failed. No retries permitted until 2025-10-10 09:25:17.058555676 +0000 UTC m=+860.074574418 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/270b1e97-6178-44d9-8d9f-ed5e4dc02000-cert") pod "infra-operator-controller-manager-656bcbd775-czldp" (UID: "270b1e97-6178-44d9-8d9f-ed5e4dc02000") : secret "infra-operator-webhook-server-cert" not found Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.562938 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.590872 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.608499 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5fvj\" (UniqueName: \"kubernetes.io/projected/270b1e97-6178-44d9-8d9f-ed5e4dc02000-kube-api-access-t5fvj\") pod \"infra-operator-controller-manager-656bcbd775-czldp\" (UID: \"270b1e97-6178-44d9-8d9f-ed5e4dc02000\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.609179 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.610508 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.618188 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5cbsb" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.625299 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlgsf\" (UniqueName: \"kubernetes.io/projected/bd3dd74a-3c58-4785-854e-400c7ba726de-kube-api-access-hlgsf\") pod \"heat-operator-controller-manager-858f76bbdd-pdnnc\" (UID: \"bd3dd74a-3c58-4785-854e-400c7ba726de\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.634357 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7528\" (UniqueName: \"kubernetes.io/projected/2132b635-c43e-4c4d-9074-37e81b6345f2-kube-api-access-w7528\") pod \"ironic-operator-controller-manager-9c5c78d49-k5nxp\" (UID: \"2132b635-c43e-4c4d-9074-37e81b6345f2\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.645093 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb6z9\" (UniqueName: \"kubernetes.io/projected/909f202d-6571-4743-9314-26cc27e94c77-kube-api-access-tb6z9\") pod \"horizon-operator-controller-manager-7ffbcb7588-96278\" (UID: \"909f202d-6571-4743-9314-26cc27e94c77\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.648934 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.662056 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrdc7\" (UniqueName: \"kubernetes.io/projected/27b6cee3-b0a3-4116-b9bb-299a81bba403-kube-api-access-mrdc7\") pod \"keystone-operator-controller-manager-55b6b7c7b8-6zprx\" (UID: \"27b6cee3-b0a3-4116-b9bb-299a81bba403\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.662128 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kr5z\" (UniqueName: \"kubernetes.io/projected/6941897d-fd2e-4ff3-ad50-aeba82eb86c2-kube-api-access-4kr5z\") pod \"manila-operator-controller-manager-5f67fbc655-q52hz\" (UID: \"6941897d-fd2e-4ff3-ad50-aeba82eb86c2\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.662149 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjsvp\" (UniqueName: \"kubernetes.io/projected/48aca3a8-684a-448f-ab17-b9d604d47484-kube-api-access-mjsvp\") pod \"nova-operator-controller-manager-5df598886f-pdf8v\" (UID: \"48aca3a8-684a-448f-ab17-b9d604d47484\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.662176 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw2t2\" (UniqueName: \"kubernetes.io/projected/bdebd762-e639-4e50-820d-cd32885a1341-kube-api-access-rw2t2\") pod \"neutron-operator-controller-manager-79d585cb66-kwn4g\" (UID: \"bdebd762-e639-4e50-820d-cd32885a1341\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.662255 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzlrf\" (UniqueName: \"kubernetes.io/projected/82353170-7f06-43e4-966b-20d0d234f0f2-kube-api-access-rzlrf\") pod \"mariadb-operator-controller-manager-f9fb45f8f-n72xm\" (UID: \"82353170-7f06-43e4-966b-20d0d234f0f2\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.707709 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.710906 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.714324 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kr5z\" (UniqueName: \"kubernetes.io/projected/6941897d-fd2e-4ff3-ad50-aeba82eb86c2-kube-api-access-4kr5z\") pod \"manila-operator-controller-manager-5f67fbc655-q52hz\" (UID: \"6941897d-fd2e-4ff3-ad50-aeba82eb86c2\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.719462 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzlrf\" (UniqueName: \"kubernetes.io/projected/82353170-7f06-43e4-966b-20d0d234f0f2-kube-api-access-rzlrf\") pod \"mariadb-operator-controller-manager-f9fb45f8f-n72xm\" (UID: \"82353170-7f06-43e4-966b-20d0d234f0f2\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.724698 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrdc7\" (UniqueName: \"kubernetes.io/projected/27b6cee3-b0a3-4116-b9bb-299a81bba403-kube-api-access-mrdc7\") pod \"keystone-operator-controller-manager-55b6b7c7b8-6zprx\" (UID: \"27b6cee3-b0a3-4116-b9bb-299a81bba403\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.732660 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.734607 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.739219 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw2t2\" (UniqueName: \"kubernetes.io/projected/bdebd762-e639-4e50-820d-cd32885a1341-kube-api-access-rw2t2\") pod \"neutron-operator-controller-manager-79d585cb66-kwn4g\" (UID: \"bdebd762-e639-4e50-820d-cd32885a1341\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.747924 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-s4nsb" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.764008 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt9jn\" (UniqueName: \"kubernetes.io/projected/43bf605f-c6c8-447b-9dc3-03be42fe7f10-kube-api-access-zt9jn\") pod \"octavia-operator-controller-manager-69fdcfc5f5-wtsqj\" (UID: \"43bf605f-c6c8-447b-9dc3-03be42fe7f10\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.764097 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjsvp\" (UniqueName: \"kubernetes.io/projected/48aca3a8-684a-448f-ab17-b9d604d47484-kube-api-access-mjsvp\") pod \"nova-operator-controller-manager-5df598886f-pdf8v\" (UID: \"48aca3a8-684a-448f-ab17-b9d604d47484\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.775375 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.777063 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.778521 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.779190 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.781784 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-27rgb" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.794137 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.794356 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9srw4" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.803944 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjsvp\" (UniqueName: \"kubernetes.io/projected/48aca3a8-684a-448f-ab17-b9d604d47484-kube-api-access-mjsvp\") pod \"nova-operator-controller-manager-5df598886f-pdf8v\" (UID: \"48aca3a8-684a-448f-ab17-b9d604d47484\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.806553 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.849576 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.857870 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.870980 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.871075 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5687f\" (UniqueName: \"kubernetes.io/projected/29661927-099c-4549-8584-9390881170d5-kube-api-access-5687f\") pod \"ovn-operator-controller-manager-79df5fb58c-r7ctx\" (UID: \"29661927-099c-4549-8584-9390881170d5\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.871119 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvx7c\" (UniqueName: \"kubernetes.io/projected/73105b8d-0527-4510-886b-7203163993fd-kube-api-access-kvx7c\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.871159 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xbn4\" (UniqueName: \"kubernetes.io/projected/db1620d5-baef-433a-b9e7-07a55004a68f-kube-api-access-8xbn4\") pod \"swift-operator-controller-manager-db6d7f97b-hjkht\" (UID: \"db1620d5-baef-433a-b9e7-07a55004a68f\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.871185 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt9jn\" (UniqueName: \"kubernetes.io/projected/43bf605f-c6c8-447b-9dc3-03be42fe7f10-kube-api-access-zt9jn\") pod \"octavia-operator-controller-manager-69fdcfc5f5-wtsqj\" (UID: \"43bf605f-c6c8-447b-9dc3-03be42fe7f10\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.894177 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.894288 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.902442 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.906751 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-tkntw" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.908544 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.909839 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.940207 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.957792 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.959378 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt9jn\" (UniqueName: \"kubernetes.io/projected/43bf605f-c6c8-447b-9dc3-03be42fe7f10-kube-api-access-zt9jn\") pod \"octavia-operator-controller-manager-69fdcfc5f5-wtsqj\" (UID: \"43bf605f-c6c8-447b-9dc3-03be42fe7f10\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.974389 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdx5j\" (UniqueName: \"kubernetes.io/projected/316d31d2-d9a9-4fd7-a079-4620ad8c5e09-kube-api-access-bdx5j\") pod \"placement-operator-controller-manager-68b6c87b68-4b4jb\" (UID: \"316d31d2-d9a9-4fd7-a079-4620ad8c5e09\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.974579 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.974737 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5687f\" (UniqueName: \"kubernetes.io/projected/29661927-099c-4549-8584-9390881170d5-kube-api-access-5687f\") pod \"ovn-operator-controller-manager-79df5fb58c-r7ctx\" (UID: \"29661927-099c-4549-8584-9390881170d5\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.974859 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvx7c\" (UniqueName: \"kubernetes.io/projected/73105b8d-0527-4510-886b-7203163993fd-kube-api-access-kvx7c\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.974963 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xbn4\" (UniqueName: \"kubernetes.io/projected/db1620d5-baef-433a-b9e7-07a55004a68f-kube-api-access-8xbn4\") pod \"swift-operator-controller-manager-db6d7f97b-hjkht\" (UID: \"db1620d5-baef-433a-b9e7-07a55004a68f\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" Oct 10 09:25:16 crc kubenswrapper[4669]: E1010 09:25:16.975468 4669 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 09:25:16 crc kubenswrapper[4669]: E1010 09:25:16.975609 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert podName:73105b8d-0527-4510-886b-7203163993fd nodeName:}" failed. No retries permitted until 2025-10-10 09:25:17.475567218 +0000 UTC m=+860.491585960 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert") pod "openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" (UID: "73105b8d-0527-4510-886b-7203163993fd") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.976201 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx"] Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.978002 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" Oct 10 09:25:16 crc kubenswrapper[4669]: I1010 09:25:16.998493 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.000251 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.004723 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.014790 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.015174 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-w429f" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.035410 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xbn4\" (UniqueName: \"kubernetes.io/projected/db1620d5-baef-433a-b9e7-07a55004a68f-kube-api-access-8xbn4\") pod \"swift-operator-controller-manager-db6d7f97b-hjkht\" (UID: \"db1620d5-baef-433a-b9e7-07a55004a68f\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.045275 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvx7c\" (UniqueName: \"kubernetes.io/projected/73105b8d-0527-4510-886b-7203163993fd-kube-api-access-kvx7c\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.062880 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.067775 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.069334 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.072125 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5687f\" (UniqueName: \"kubernetes.io/projected/29661927-099c-4549-8584-9390881170d5-kube-api-access-5687f\") pod \"ovn-operator-controller-manager-79df5fb58c-r7ctx\" (UID: \"29661927-099c-4549-8584-9390881170d5\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.076123 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4w7k\" (UniqueName: \"kubernetes.io/projected/f30a8280-3312-4333-8966-810b8dd3c889-kube-api-access-t4w7k\") pod \"telemetry-operator-controller-manager-67cfc6749b-j7fm7\" (UID: \"f30a8280-3312-4333-8966-810b8dd3c889\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.076158 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/270b1e97-6178-44d9-8d9f-ed5e4dc02000-cert\") pod \"infra-operator-controller-manager-656bcbd775-czldp\" (UID: \"270b1e97-6178-44d9-8d9f-ed5e4dc02000\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.076225 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdx5j\" (UniqueName: \"kubernetes.io/projected/316d31d2-d9a9-4fd7-a079-4620ad8c5e09-kube-api-access-bdx5j\") pod \"placement-operator-controller-manager-68b6c87b68-4b4jb\" (UID: \"316d31d2-d9a9-4fd7-a079-4620ad8c5e09\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.080800 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-gl54t" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.082679 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.096513 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/270b1e97-6178-44d9-8d9f-ed5e4dc02000-cert\") pod \"infra-operator-controller-manager-656bcbd775-czldp\" (UID: \"270b1e97-6178-44d9-8d9f-ed5e4dc02000\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.107664 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.108816 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.110493 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.128320 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sldlg" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.143460 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdx5j\" (UniqueName: \"kubernetes.io/projected/316d31d2-d9a9-4fd7-a079-4620ad8c5e09-kube-api-access-bdx5j\") pod \"placement-operator-controller-manager-68b6c87b68-4b4jb\" (UID: \"316d31d2-d9a9-4fd7-a079-4620ad8c5e09\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.177218 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4w7k\" (UniqueName: \"kubernetes.io/projected/f30a8280-3312-4333-8966-810b8dd3c889-kube-api-access-t4w7k\") pod \"telemetry-operator-controller-manager-67cfc6749b-j7fm7\" (UID: \"f30a8280-3312-4333-8966-810b8dd3c889\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.177282 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6klws\" (UniqueName: \"kubernetes.io/projected/5d2f190a-4eae-44e6-a5cb-e15abf2e037d-kube-api-access-6klws\") pod \"watcher-operator-controller-manager-7f554bff7b-cjtjr\" (UID: \"5d2f190a-4eae-44e6-a5cb-e15abf2e037d\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.177323 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r68kg\" (UniqueName: \"kubernetes.io/projected/684c8731-c863-49d2-a029-d9d69ae01f8d-kube-api-access-r68kg\") pod \"test-operator-controller-manager-5458f77c4-b2p7l\" (UID: \"684c8731-c863-49d2-a029-d9d69ae01f8d\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.178664 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.258534 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4w7k\" (UniqueName: \"kubernetes.io/projected/f30a8280-3312-4333-8966-810b8dd3c889-kube-api-access-t4w7k\") pod \"telemetry-operator-controller-manager-67cfc6749b-j7fm7\" (UID: \"f30a8280-3312-4333-8966-810b8dd3c889\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.259786 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.280846 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r68kg\" (UniqueName: \"kubernetes.io/projected/684c8731-c863-49d2-a029-d9d69ae01f8d-kube-api-access-r68kg\") pod \"test-operator-controller-manager-5458f77c4-b2p7l\" (UID: \"684c8731-c863-49d2-a029-d9d69ae01f8d\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.280986 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6klws\" (UniqueName: \"kubernetes.io/projected/5d2f190a-4eae-44e6-a5cb-e15abf2e037d-kube-api-access-6klws\") pod \"watcher-operator-controller-manager-7f554bff7b-cjtjr\" (UID: \"5d2f190a-4eae-44e6-a5cb-e15abf2e037d\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.282708 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.320734 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r68kg\" (UniqueName: \"kubernetes.io/projected/684c8731-c863-49d2-a029-d9d69ae01f8d-kube-api-access-r68kg\") pod \"test-operator-controller-manager-5458f77c4-b2p7l\" (UID: \"684c8731-c863-49d2-a029-d9d69ae01f8d\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.340452 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.341506 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.348211 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.351881 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.352035 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-7mgls" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.362059 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.372494 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6klws\" (UniqueName: \"kubernetes.io/projected/5d2f190a-4eae-44e6-a5cb-e15abf2e037d-kube-api-access-6klws\") pod \"watcher-operator-controller-manager-7f554bff7b-cjtjr\" (UID: \"5d2f190a-4eae-44e6-a5cb-e15abf2e037d\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.382028 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.400707 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.422942 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.430613 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.443184 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-n6jtd" Oct 10 09:25:17 crc kubenswrapper[4669]: W1010 09:25:17.465790 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dcc64bf_ac2d_4357_adcb_b6c146462464.slice/crio-1ff7ded3837248274e4ad3922a0e3be44420d1ad85d62d17c31052bbd302e00c WatchSource:0}: Error finding container 1ff7ded3837248274e4ad3922a0e3be44420d1ad85d62d17c31052bbd302e00c: Status 404 returned error can't find the container with id 1ff7ded3837248274e4ad3922a0e3be44420d1ad85d62d17c31052bbd302e00c Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.477255 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.484158 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.484192 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hvh8\" (UniqueName: \"kubernetes.io/projected/5f2708ab-b3ba-4556-9e11-7018b186975e-kube-api-access-6hvh8\") pod \"openstack-operator-controller-manager-5698bb9464-flrhj\" (UID: \"5f2708ab-b3ba-4556-9e11-7018b186975e\") " pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.484236 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8mbh\" (UniqueName: \"kubernetes.io/projected/637efcba-83b3-4f3a-93f0-1ce6257987e1-kube-api-access-g8mbh\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs\" (UID: \"637efcba-83b3-4f3a-93f0-1ce6257987e1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.484261 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5f2708ab-b3ba-4556-9e11-7018b186975e-cert\") pod \"openstack-operator-controller-manager-5698bb9464-flrhj\" (UID: \"5f2708ab-b3ba-4556-9e11-7018b186975e\") " pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:17 crc kubenswrapper[4669]: E1010 09:25:17.484278 4669 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 09:25:17 crc kubenswrapper[4669]: E1010 09:25:17.484315 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert podName:73105b8d-0527-4510-886b-7203163993fd nodeName:}" failed. No retries permitted until 2025-10-10 09:25:18.484302472 +0000 UTC m=+861.500321214 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert") pod "openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" (UID: "73105b8d-0527-4510-886b-7203163993fd") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.518985 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.531890 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.605812 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8mbh\" (UniqueName: \"kubernetes.io/projected/637efcba-83b3-4f3a-93f0-1ce6257987e1-kube-api-access-g8mbh\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs\" (UID: \"637efcba-83b3-4f3a-93f0-1ce6257987e1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.605857 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5f2708ab-b3ba-4556-9e11-7018b186975e-cert\") pod \"openstack-operator-controller-manager-5698bb9464-flrhj\" (UID: \"5f2708ab-b3ba-4556-9e11-7018b186975e\") " pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.605942 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hvh8\" (UniqueName: \"kubernetes.io/projected/5f2708ab-b3ba-4556-9e11-7018b186975e-kube-api-access-6hvh8\") pod \"openstack-operator-controller-manager-5698bb9464-flrhj\" (UID: \"5f2708ab-b3ba-4556-9e11-7018b186975e\") " pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:17 crc kubenswrapper[4669]: E1010 09:25:17.606488 4669 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 10 09:25:17 crc kubenswrapper[4669]: E1010 09:25:17.606541 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f2708ab-b3ba-4556-9e11-7018b186975e-cert podName:5f2708ab-b3ba-4556-9e11-7018b186975e nodeName:}" failed. No retries permitted until 2025-10-10 09:25:18.106524495 +0000 UTC m=+861.122543237 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5f2708ab-b3ba-4556-9e11-7018b186975e-cert") pod "openstack-operator-controller-manager-5698bb9464-flrhj" (UID: "5f2708ab-b3ba-4556-9e11-7018b186975e") : secret "webhook-server-cert" not found Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.647010 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hvh8\" (UniqueName: \"kubernetes.io/projected/5f2708ab-b3ba-4556-9e11-7018b186975e-kube-api-access-6hvh8\") pod \"openstack-operator-controller-manager-5698bb9464-flrhj\" (UID: \"5f2708ab-b3ba-4556-9e11-7018b186975e\") " pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.648805 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8mbh\" (UniqueName: \"kubernetes.io/projected/637efcba-83b3-4f3a-93f0-1ce6257987e1-kube-api-access-g8mbh\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs\" (UID: \"637efcba-83b3-4f3a-93f0-1ce6257987e1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.666639 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.728414 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n"] Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.885826 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" Oct 10 09:25:17 crc kubenswrapper[4669]: I1010 09:25:17.972177 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v"] Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.051207 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" event={"ID":"7dcc64bf-ac2d-4357-adcb-b6c146462464","Type":"ContainerStarted","Data":"1ff7ded3837248274e4ad3922a0e3be44420d1ad85d62d17c31052bbd302e00c"} Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.079748 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" event={"ID":"55f7e993-3dd8-4acf-9716-47cef7f73ef8","Type":"ContainerStarted","Data":"f138813eb205fca2cc2ac29b6f8118e11e60f29f38deb98bb3640f24e1167731"} Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.095501 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wq94g" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="registry-server" containerID="cri-o://9d1fcd2ed9f0e77f9e5243cde468c6399dcd7a6b506df4e2532297d0ecb1b77a" gracePeriod=2 Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.095647 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" event={"ID":"1a5e08a1-2d18-4382-8129-be1376c40bde","Type":"ContainerStarted","Data":"6b192c997a59a33880996e17080f274cc83b72efbf05dbe2579d6143dfb603af"} Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.109989 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc"] Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.153355 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5f2708ab-b3ba-4556-9e11-7018b186975e-cert\") pod \"openstack-operator-controller-manager-5698bb9464-flrhj\" (UID: \"5f2708ab-b3ba-4556-9e11-7018b186975e\") " pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.185212 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5f2708ab-b3ba-4556-9e11-7018b186975e-cert\") pod \"openstack-operator-controller-manager-5698bb9464-flrhj\" (UID: \"5f2708ab-b3ba-4556-9e11-7018b186975e\") " pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.265396 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.413938 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v"] Oct 10 09:25:18 crc kubenswrapper[4669]: W1010 09:25:18.464913 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48aca3a8_684a_448f_ab17_b9d604d47484.slice/crio-f5d2bb6cf295986480db3d36371b2534145fb2a9d868a5779d6156fd11e2982f WatchSource:0}: Error finding container f5d2bb6cf295986480db3d36371b2534145fb2a9d868a5779d6156fd11e2982f: Status 404 returned error can't find the container with id f5d2bb6cf295986480db3d36371b2534145fb2a9d868a5779d6156fd11e2982f Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.565551 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.566850 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx"] Oct 10 09:25:18 crc kubenswrapper[4669]: W1010 09:25:18.607129 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27b6cee3_b0a3_4116_b9bb_299a81bba403.slice/crio-2ed4f64227e21a1dc9319681e53d9c89b1220ed34f19416245594b363266a299 WatchSource:0}: Error finding container 2ed4f64227e21a1dc9319681e53d9c89b1220ed34f19416245594b363266a299: Status 404 returned error can't find the container with id 2ed4f64227e21a1dc9319681e53d9c89b1220ed34f19416245594b363266a299 Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.608366 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73105b8d-0527-4510-886b-7203163993fd-cert\") pod \"openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv\" (UID: \"73105b8d-0527-4510-886b-7203163993fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.639206 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm"] Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.664603 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz"] Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.675689 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.691145 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278"] Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.703763 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g"] Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.741666 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj"] Oct 10 09:25:18 crc kubenswrapper[4669]: W1010 09:25:18.755523 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43bf605f_c6c8_447b_9dc3_03be42fe7f10.slice/crio-36a7dc3b1f2c1ed34e6217340ce25c9c144c9ef66af4717b42c1d14e00c9ba3a WatchSource:0}: Error finding container 36a7dc3b1f2c1ed34e6217340ce25c9c144c9ef66af4717b42c1d14e00c9ba3a: Status 404 returned error can't find the container with id 36a7dc3b1f2c1ed34e6217340ce25c9c144c9ef66af4717b42c1d14e00c9ba3a Oct 10 09:25:18 crc kubenswrapper[4669]: I1010 09:25:18.766573 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp"] Oct 10 09:25:18 crc kubenswrapper[4669]: W1010 09:25:18.792706 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2132b635_c43e_4c4d_9074_37e81b6345f2.slice/crio-eb3bac287c72a4c320fe419341cee10141749ea91d4d40bc39c7d5d6850f4056 WatchSource:0}: Error finding container eb3bac287c72a4c320fe419341cee10141749ea91d4d40bc39c7d5d6850f4056: Status 404 returned error can't find the container with id eb3bac287c72a4c320fe419341cee10141749ea91d4d40bc39c7d5d6850f4056 Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.110818 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" event={"ID":"43bf605f-c6c8-447b-9dc3-03be42fe7f10","Type":"ContainerStarted","Data":"36a7dc3b1f2c1ed34e6217340ce25c9c144c9ef66af4717b42c1d14e00c9ba3a"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.112016 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" event={"ID":"6941897d-fd2e-4ff3-ad50-aeba82eb86c2","Type":"ContainerStarted","Data":"5b16d293dc8f66c71372a5b01a77cb3757b79c664558b4c0e789850b6140c056"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.112872 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" event={"ID":"909f202d-6571-4743-9314-26cc27e94c77","Type":"ContainerStarted","Data":"e81dc28a8178a980cce97874dde8f3dcb78b42be7a37c2594b829ddd9398bb90"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.114500 4669 generic.go:334] "Generic (PLEG): container finished" podID="5c07ede9-cc73-43c5-812b-0db964c66349" containerID="9d1fcd2ed9f0e77f9e5243cde468c6399dcd7a6b506df4e2532297d0ecb1b77a" exitCode=0 Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.114543 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq94g" event={"ID":"5c07ede9-cc73-43c5-812b-0db964c66349","Type":"ContainerDied","Data":"9d1fcd2ed9f0e77f9e5243cde468c6399dcd7a6b506df4e2532297d0ecb1b77a"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.123069 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" event={"ID":"bd3dd74a-3c58-4785-854e-400c7ba726de","Type":"ContainerStarted","Data":"51424b2c9978fe5a78efc52eb16bef752f5e484a9538ac8f4bed4e3a59fae0b0"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.130834 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" event={"ID":"82353170-7f06-43e4-966b-20d0d234f0f2","Type":"ContainerStarted","Data":"2fbb25ba700dfd501315512b4c769fa4328de7d8fed5409686d849dddcffb45e"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.137715 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" event={"ID":"2132b635-c43e-4c4d-9074-37e81b6345f2","Type":"ContainerStarted","Data":"eb3bac287c72a4c320fe419341cee10141749ea91d4d40bc39c7d5d6850f4056"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.146722 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" event={"ID":"48aca3a8-684a-448f-ab17-b9d604d47484","Type":"ContainerStarted","Data":"f5d2bb6cf295986480db3d36371b2534145fb2a9d868a5779d6156fd11e2982f"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.156792 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" event={"ID":"bdebd762-e639-4e50-820d-cd32885a1341","Type":"ContainerStarted","Data":"7a87d46f57d073cf82275682b0066586dcabc1212314c781873f789a003f1bb9"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.167784 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" event={"ID":"affd00ea-eac5-4c5f-b452-c8bb0e17aba8","Type":"ContainerStarted","Data":"5c331c8b3fe31c3f11049647a18116d0ce471346618f1d16299ec0e4006df3b2"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.184137 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" event={"ID":"27b6cee3-b0a3-4116-b9bb-299a81bba403","Type":"ContainerStarted","Data":"2ed4f64227e21a1dc9319681e53d9c89b1220ed34f19416245594b363266a299"} Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.215540 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.230100 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.270450 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx"] Oct 10 09:25:19 crc kubenswrapper[4669]: W1010 09:25:19.333061 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29661927_099c_4549_8584_9390881170d5.slice/crio-abe3637abc2e4cc43a87826ef03d0fc24a31264f4caad7f951c3815f32de1895 WatchSource:0}: Error finding container abe3637abc2e4cc43a87826ef03d0fc24a31264f4caad7f951c3815f32de1895: Status 404 returned error can't find the container with id abe3637abc2e4cc43a87826ef03d0fc24a31264f4caad7f951c3815f32de1895 Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.341835 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.356687 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-czldp"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.372554 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.390885 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.440537 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.466813 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.513208 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv"] Oct 10 09:25:19 crc kubenswrapper[4669]: E1010 09:25:19.516959 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6klws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7f554bff7b-cjtjr_openstack-operators(5d2f190a-4eae-44e6-a5cb-e15abf2e037d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 09:25:19 crc kubenswrapper[4669]: W1010 09:25:19.539387 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb1620d5_baef_433a_b9e7_07a55004a68f.slice/crio-68bca58665f05e77325c849f6eb2e52f90a4fb91444390688b11ba20ee269eab WatchSource:0}: Error finding container 68bca58665f05e77325c849f6eb2e52f90a4fb91444390688b11ba20ee269eab: Status 404 returned error can't find the container with id 68bca58665f05e77325c849f6eb2e52f90a4fb91444390688b11ba20ee269eab Oct 10 09:25:19 crc kubenswrapper[4669]: W1010 09:25:19.553882 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73105b8d_0527_4510_886b_7203163993fd.slice/crio-0087af5012001dbcf675bd1f0361ecdd193c98c1f9f6ad5bb3a2784c3a6070dd WatchSource:0}: Error finding container 0087af5012001dbcf675bd1f0361ecdd193c98c1f9f6ad5bb3a2784c3a6070dd: Status 404 returned error can't find the container with id 0087af5012001dbcf675bd1f0361ecdd193c98c1f9f6ad5bb3a2784c3a6070dd Oct 10 09:25:19 crc kubenswrapper[4669]: E1010 09:25:19.614969 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:03b4f3db4b373515f7e4095984b97197c05a14f87b2a0a525eb5d7be1d7bda66,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:6722a752fb7cbffbae811f6ad6567120fbd4ebbe8c38a83ec2df02850a3276bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:2115452234aedb505ed4efc6cd9b9a4ce3b9809aa7d0128d8fbeeee84dad1a69,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:50597a8eaa6c4383f357574dcab8358b698729797b4156d932985a08ab86b7cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:cb4997d62c7b2534233a676cb92e19cf85dda07e2fb9fa642c28aab30489f69a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:1ccbf3f6cf24c9ee91bed71467491e22b8cb4b95bce90250f4174fae936b0fa1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:e7dcc3bf23d5e0393ac173e3c43d4ae85f4613a4fd16b3c147dc32ae491d49bf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:2a1a8b582c6e4cc31081bd8b0887acf45e31c1d14596c4e361d27d08fef0debf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:86daeb9c834bfcedb533086dff59a6b5b6e832b94ce2a9116337f8736bb80032,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:5d4fdf424fad33a3650163e9e7423f92e97de3305508c2b7c6435822e0313189,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:6d28de018f6e1672e775a75735e3bc16b63da41acd8fb5196ee0b06856c07133,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:c5fc9b72fc593bcf3b569c7ed24a256448eb1afab1504e668a3822e978be1306,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:88b99249f15470f359fb554f7f3a56974b743f4655e3f0c982c0260f75a67697,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:e861d66785047d39eb68d9bac23e3f57ac84d9bd95593502d9b3b913b99fd1a4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:b95f09bf3d259f9eacf3b63931977483f5c3c332f49b95ee8a69d8e3fb71d082,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:6fc7801c0d18d41b9f11484b1cdb342de9cebd93072ec2205dbe40945715184f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:d4d824b80cbed683543d9e8c7045ac97e080774f45a5067ccbca26404e067821,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:182ec75938d8d3fb7d8f916373368add24062fec90489aa57776a81d0b36ea20,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:9507ba5ab74cbae902e2dc07f89c7b3b5b76d8079e444365fe0eee6000fd7aaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:17db080dcc4099f8a20aa0f238b6bca5c104672ae46743adeab9d1637725ecaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:fd55cf3d73bfdc518419c9ba0b0cbef275140ae2d3bd0342a7310f81d57c2d78,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:d164a9bd383f50df69fc22e7422f4650cd5076c90ed19278fc0f04e54345a63d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:6beffe7d0bd75f9d1f495aeb7ab2334a2414af2c581d4833363df8441ed01018,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:261e76f60c6bc6b172dc3608504552c63e83358a4fa3c0952a671544d83aa83f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2308c7b6c3d0aabbadfc9a06d84d67d2243f27fe8eed740ee96b1ce910203f62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:02d33f59749441cd5751c319e9d7cff97ab1004844c0e992650d340c6e8fbf43,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:9cf0ca292340f1f978603955ef682effbf24316d6e2376b1c89906d84c3f06d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:c3e651f35b930bcf1a3084be8910c2f3f34d22a976c5379cf518a68d9994bfa7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:58f678016d7f6c8fe579abe886fd138ef853642faa6766ca60639feac12d82ac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:46f92909153aaf03a585374b77d103c536509747e3270558d9a533295c46a7c5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:7fe367f51638c5c302fd3f8e66a31b09cb3b11519a7f72ef142b6c6fe8b91694,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:9ebf424d4107275a2e3f21f7a18ef257ff2f97c1298109ac7c802a5a4f4794f2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:4fcbe0d9a3c845708ecc32102ad4abbcbd947d87e5cf91f186de75b5d84ec681,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:58a4e9a4dea86635c93ce37a2bb3c60ece62b3d656f6ee6a8845347cbb3e90fd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:6f2b843bc9f4ceb1ee873972d69e6bae6e1dbd378b486995bc3697d8bcff6339,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:7323406a63fb3fdbb3eea4da0f7e8ed89c94c9bd0ad5ecd6c18fa4a4c2c550c4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:7ae82068011e2d2e5ddc88c943fd32ff4a11902793e7a1df729811b2e27122a0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:0c762c15d9d98d39cc9dc3d1f9a70f9188fef58d4e2f3b0c69c896cab8da5e48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:febf65561eeef5b36b70d0d65ee83f6451e43ec97bfab4d826e14215da6ff19b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:b8aadfc3d547c5ef1e27fcb573d4760cf8c2f2271eefe1793c35a0d46b640837,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:ecc91fd5079ee6d0c6ae1b11e97da790e33864d0e1930e574f959da2bddfa59a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:2e981e93f99c929a3f04e5e41c8f645d44d390a9aeee3c5193cce7ec2edcbf3a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:1e5714637b6e1a24c2858fe6d9bbb3f00bc61d69ad74a657b1c23682bf4cb2b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:35b8dcf27dc3b67f3840fa0e693ff312f74f7e22c634dff206a5c4d0133c716c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:e109e4863e05e803dbfe04917756fd52231c560c65353170a2000be6cc2bb53d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:6df0bebd9318ce11624413249e7e9781311638f276f8877668d3b382fe90e62f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:a51ed62767206067aa501142dbf01f20b3d65325d30faf1b4d6424d5b17dfba5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:592e3cd32d3cc97a69093ad905b449aa374ffbb1b2644b738bb6c1434476d1f6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:5f179b847f2dc32d9110b8f2be9fe65f1aeada1e18105dffdaf052981215d844,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:9596452e283febbe08204d0ef0fd1992af3395d0969f7ac76663ed7c8be5b4d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:d61005a10bef1b37762a8a41e6755c1169241e36cc5f92886bca6f4f6b9c381a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:e6a4335bcbeed3cd3e73ac879f754e314761e4a417a67539ca88e96a79346328,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:97d88fc53421b699fc91983313d7beec4a0f177089e95bdf5ba15c3f521db9a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:5365e5c9c3ad2ede1b6945255b2cc6b009d642c39babdf25e0655282cfa646fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:5b55795d774e0ea160ff8a7fd491ed41cf2d93c7d821694abb3a879eaffcefeb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:26e955c46a6063eafcfeb79430bf3d9268dbe95687c00e63a624b3ec5a846f5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:58939baa18ab09e2b24996c5f3665ae52274b781f661ea06a67c991e9a832d5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:b8bff6857fec93c3c1521f1a8c23de21bcb86fc0f960972e81f6c3f95d4185be,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:943eee724277e252795909137538a553ef5284c8103ad01b9be7b0138c66d14d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:d97b08fd421065c8c33a523973822ac468500cbe853069aa9214393fbda7a908,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:d76f7d6620930cc2e9ac070492bbeb525f83ce5ff4947463e3784bf1ce04a857,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:289dea3beea1cd4405895fc42e44372b35e4a941e31c59e102c333471a3ca9b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:9b19894fa67a81bf8ba4159b55b49f38877c670aeb97e2021c341cef2a9294e4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:ea164961ad30453ad0301c6b73364e1f1024f689634c88dd98265f9c7048e31d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:6f9f2ea45f0271f6da8eb05a5f74cf5ce6769479346f5c2f407ee6f31a9c7ff3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:59448516174fc3bab679b9a8dd62cb9a9d16b5734aadbeb98e960e3b7c79bd22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:2bf32d9b95899d7637dfe19d07cf1ecc9a06593984faff57a3c0dce060012edb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7a452cd18b64d522e8a1e25bdcea543e9fe5f5b76e1c5e044c2b5334e06a326b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:6a46aa13aa359b8e782a22d67db42db02bbf2bb7e35df4b684ac1daeda38cde3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:f6824854bea6b2acbb00c34639799b4744818d4adbdd40e37dc5088f9ae18d58,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:a66d2fdc21f25c690f02e643d2666dbe7df43a64cd55086ec33d6755e6d809b9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:ad5cec8b914687f3b378754f76bd30ade09c1b33a5638816b64fee68ebe2ab45,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:d04912a69e111cb6ca00b5019cdc2ebc43b89e5fc090260718752184a2581072,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:f9a32c333aae6ef5bddd7ba613c17d42207d290e58c079b80235621fe2cd626c,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kvx7c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv_openstack-operators(73105b8d-0527-4510-886b-7203163993fd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.838652 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8fhcl"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.857119 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8fhcl"] Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.857257 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:19 crc kubenswrapper[4669]: E1010 09:25:19.891916 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" podUID="5d2f190a-4eae-44e6-a5cb-e15abf2e037d" Oct 10 09:25:19 crc kubenswrapper[4669]: I1010 09:25:19.993426 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.011302 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nvxr\" (UniqueName: \"kubernetes.io/projected/178baabb-2525-43c3-b1c3-844f4311f786-kube-api-access-4nvxr\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.011398 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-utilities\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.011449 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-catalog-content\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.111954 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-catalog-content\") pod \"5c07ede9-cc73-43c5-812b-0db964c66349\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.112037 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-utilities\") pod \"5c07ede9-cc73-43c5-812b-0db964c66349\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.112073 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp98x\" (UniqueName: \"kubernetes.io/projected/5c07ede9-cc73-43c5-812b-0db964c66349-kube-api-access-pp98x\") pod \"5c07ede9-cc73-43c5-812b-0db964c66349\" (UID: \"5c07ede9-cc73-43c5-812b-0db964c66349\") " Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.112307 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-catalog-content\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.112368 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nvxr\" (UniqueName: \"kubernetes.io/projected/178baabb-2525-43c3-b1c3-844f4311f786-kube-api-access-4nvxr\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.112432 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-utilities\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.113224 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-utilities" (OuterVolumeSpecName: "utilities") pod "5c07ede9-cc73-43c5-812b-0db964c66349" (UID: "5c07ede9-cc73-43c5-812b-0db964c66349"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.113400 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-catalog-content\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.113412 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-utilities\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.131653 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c07ede9-cc73-43c5-812b-0db964c66349-kube-api-access-pp98x" (OuterVolumeSpecName: "kube-api-access-pp98x") pod "5c07ede9-cc73-43c5-812b-0db964c66349" (UID: "5c07ede9-cc73-43c5-812b-0db964c66349"). InnerVolumeSpecName "kube-api-access-pp98x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.132990 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nvxr\" (UniqueName: \"kubernetes.io/projected/178baabb-2525-43c3-b1c3-844f4311f786-kube-api-access-4nvxr\") pod \"redhat-operators-8fhcl\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: E1010 09:25:20.157440 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" podUID="73105b8d-0527-4510-886b-7203163993fd" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.211536 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.213300 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.213375 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp98x\" (UniqueName: \"kubernetes.io/projected/5c07ede9-cc73-43c5-812b-0db964c66349-kube-api-access-pp98x\") on node \"crc\" DevicePath \"\"" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.214703 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c07ede9-cc73-43c5-812b-0db964c66349" (UID: "5c07ede9-cc73-43c5-812b-0db964c66349"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.216072 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" event={"ID":"5d2f190a-4eae-44e6-a5cb-e15abf2e037d","Type":"ContainerStarted","Data":"443ab251511725babf6fb128f8c528d55fcea3c2592a9764064e957844ce351d"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.216397 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" event={"ID":"5d2f190a-4eae-44e6-a5cb-e15abf2e037d","Type":"ContainerStarted","Data":"693cddf49c42705a991d083a773e108ea8afe0bf35697c3fcd99442cd6998f10"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.228419 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" event={"ID":"5f2708ab-b3ba-4556-9e11-7018b186975e","Type":"ContainerStarted","Data":"4adaf12606ff7f1c8354f5d3ab9930ae06b3f29d653f08ef03ce931ad79f6f73"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.228482 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" event={"ID":"5f2708ab-b3ba-4556-9e11-7018b186975e","Type":"ContainerStarted","Data":"cf243d96db4d694eb9b77639dd042c837f3d7adc26adfaaa87d0a8caba8bcd4f"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.230672 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" event={"ID":"270b1e97-6178-44d9-8d9f-ed5e4dc02000","Type":"ContainerStarted","Data":"be564964768cd9fbfa4bb9fef8904d134c610724c9d95d361e3b43efc31297ed"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.232483 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" event={"ID":"637efcba-83b3-4f3a-93f0-1ce6257987e1","Type":"ContainerStarted","Data":"236f0b3fa01380a4d0bfd990833b55377c6c6f7540d5adcb7b1c46bca20e0161"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.239531 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" event={"ID":"684c8731-c863-49d2-a029-d9d69ae01f8d","Type":"ContainerStarted","Data":"9162f4d69a72bd7eed1eebed4304e442ae4b5a0fa27ee2fe5c4c9a4176f9b5e3"} Oct 10 09:25:20 crc kubenswrapper[4669]: E1010 09:25:20.244065 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" podUID="5d2f190a-4eae-44e6-a5cb-e15abf2e037d" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.245474 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" event={"ID":"316d31d2-d9a9-4fd7-a079-4620ad8c5e09","Type":"ContainerStarted","Data":"e74d4459505b902216173724d74535455f98d8110636a0a3e9889849c54b2f5f"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.252005 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" event={"ID":"f30a8280-3312-4333-8966-810b8dd3c889","Type":"ContainerStarted","Data":"838f38fe219a5767a6a0871f551272eb1fc1b604b4235a636e37411b3d44af6a"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.262601 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq94g" event={"ID":"5c07ede9-cc73-43c5-812b-0db964c66349","Type":"ContainerDied","Data":"8949dcf0569735706770c8a1fa66cacf3c83356b14d80e3793f6a4eeb272de07"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.262656 4669 scope.go:117] "RemoveContainer" containerID="9d1fcd2ed9f0e77f9e5243cde468c6399dcd7a6b506df4e2532297d0ecb1b77a" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.262718 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq94g" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.265628 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" event={"ID":"29661927-099c-4549-8584-9390881170d5","Type":"ContainerStarted","Data":"abe3637abc2e4cc43a87826ef03d0fc24a31264f4caad7f951c3815f32de1895"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.270896 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" event={"ID":"db1620d5-baef-433a-b9e7-07a55004a68f","Type":"ContainerStarted","Data":"68bca58665f05e77325c849f6eb2e52f90a4fb91444390688b11ba20ee269eab"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.292549 4669 scope.go:117] "RemoveContainer" containerID="a15a2ee89bb8f9640d1ca743a3e3b21e4f7e7eaabfb7bf83b2bec4bf3eaed075" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.303647 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" event={"ID":"73105b8d-0527-4510-886b-7203163993fd","Type":"ContainerStarted","Data":"dc0f121558c745feadd0cbf05108af1d1b1b828f777aa2a2c3fd1aa9c5165c98"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.303722 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" event={"ID":"73105b8d-0527-4510-886b-7203163993fd","Type":"ContainerStarted","Data":"0087af5012001dbcf675bd1f0361ecdd193c98c1f9f6ad5bb3a2784c3a6070dd"} Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.305903 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wq94g"] Oct 10 09:25:20 crc kubenswrapper[4669]: E1010 09:25:20.309288 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" podUID="73105b8d-0527-4510-886b-7203163993fd" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.314924 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c07ede9-cc73-43c5-812b-0db964c66349-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.319054 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wq94g"] Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.456089 4669 scope.go:117] "RemoveContainer" containerID="f16b6d40f7eb99af350d77b87666dff9944fc10a492ea30dffec028e3fa0cecb" Oct 10 09:25:20 crc kubenswrapper[4669]: I1010 09:25:20.831096 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8fhcl"] Oct 10 09:25:21 crc kubenswrapper[4669]: I1010 09:25:21.324396 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" event={"ID":"5f2708ab-b3ba-4556-9e11-7018b186975e","Type":"ContainerStarted","Data":"ccd0fa921d2168b1a28e43a6543fe5a1701c5afc6bb80bc708b8350145c40206"} Oct 10 09:25:21 crc kubenswrapper[4669]: I1010 09:25:21.325580 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:21 crc kubenswrapper[4669]: I1010 09:25:21.334171 4669 generic.go:334] "Generic (PLEG): container finished" podID="178baabb-2525-43c3-b1c3-844f4311f786" containerID="3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517" exitCode=0 Oct 10 09:25:21 crc kubenswrapper[4669]: I1010 09:25:21.334232 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fhcl" event={"ID":"178baabb-2525-43c3-b1c3-844f4311f786","Type":"ContainerDied","Data":"3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517"} Oct 10 09:25:21 crc kubenswrapper[4669]: I1010 09:25:21.334257 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fhcl" event={"ID":"178baabb-2525-43c3-b1c3-844f4311f786","Type":"ContainerStarted","Data":"18d98314612b7b3d7ed94a7edbfef7897d25367bc1a6e90a571bdcf73f5260cf"} Oct 10 09:25:21 crc kubenswrapper[4669]: E1010 09:25:21.350419 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" podUID="5d2f190a-4eae-44e6-a5cb-e15abf2e037d" Oct 10 09:25:21 crc kubenswrapper[4669]: I1010 09:25:21.357790 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" podStartSLOduration=4.357774218 podStartE2EDuration="4.357774218s" podCreationTimestamp="2025-10-10 09:25:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:25:21.351005757 +0000 UTC m=+864.367024499" watchObservedRunningTime="2025-10-10 09:25:21.357774218 +0000 UTC m=+864.373792960" Oct 10 09:25:21 crc kubenswrapper[4669]: E1010 09:25:21.365237 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" podUID="73105b8d-0527-4510-886b-7203163993fd" Oct 10 09:25:21 crc kubenswrapper[4669]: I1010 09:25:21.846669 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" path="/var/lib/kubelet/pods/5c07ede9-cc73-43c5-812b-0db964c66349/volumes" Oct 10 09:25:23 crc kubenswrapper[4669]: I1010 09:25:23.391177 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fhcl" event={"ID":"178baabb-2525-43c3-b1c3-844f4311f786","Type":"ContainerStarted","Data":"97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117"} Oct 10 09:25:26 crc kubenswrapper[4669]: I1010 09:25:26.417932 4669 generic.go:334] "Generic (PLEG): container finished" podID="178baabb-2525-43c3-b1c3-844f4311f786" containerID="97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117" exitCode=0 Oct 10 09:25:26 crc kubenswrapper[4669]: I1010 09:25:26.418008 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fhcl" event={"ID":"178baabb-2525-43c3-b1c3-844f4311f786","Type":"ContainerDied","Data":"97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117"} Oct 10 09:25:28 crc kubenswrapper[4669]: I1010 09:25:28.271167 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5698bb9464-flrhj" Oct 10 09:25:33 crc kubenswrapper[4669]: E1010 09:25:33.386994 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577" Oct 10 09:25:33 crc kubenswrapper[4669]: E1010 09:25:33.387661 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k4swb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-658bdf4b74-sbc5v_openstack-operators(affd00ea-eac5-4c5f-b452-c8bb0e17aba8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:33 crc kubenswrapper[4669]: E1010 09:25:33.779129 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a" Oct 10 09:25:33 crc kubenswrapper[4669]: E1010 09:25:33.779340 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r68kg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5458f77c4-b2p7l_openstack-operators(684c8731-c863-49d2-a029-d9d69ae01f8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:35 crc kubenswrapper[4669]: E1010 09:25:35.579606 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0" Oct 10 09:25:35 crc kubenswrapper[4669]: E1010 09:25:35.580305 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t4w7k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-67cfc6749b-j7fm7_openstack-operators(f30a8280-3312-4333-8966-810b8dd3c889): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:35 crc kubenswrapper[4669]: E1010 09:25:35.974881 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:73736f216f886549901fbcfc823b072f73691c9a79ec79e59d100e992b9c1e34" Oct 10 09:25:35 crc kubenswrapper[4669]: E1010 09:25:35.975062 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:73736f216f886549901fbcfc823b072f73691c9a79ec79e59d100e992b9c1e34,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nqw45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-85d5d9dd78-pn8xd_openstack-operators(1a5e08a1-2d18-4382-8129-be1376c40bde): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:36 crc kubenswrapper[4669]: E1010 09:25:36.377874 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7" Oct 10 09:25:36 crc kubenswrapper[4669]: E1010 09:25:36.378117 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:09deecf840d38ff6af3c924729cf0a9444bc985848bfbe7c918019b88a6bc4d7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zt9jn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69fdcfc5f5-wtsqj_openstack-operators(43bf605f-c6c8-447b-9dc3-03be42fe7f10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:37 crc kubenswrapper[4669]: E1010 09:25:37.712842 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997" Oct 10 09:25:37 crc kubenswrapper[4669]: E1010 09:25:37.714288 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mrdc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-55b6b7c7b8-6zprx_openstack-operators(27b6cee3-b0a3-4116-b9bb-299a81bba403): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:38 crc kubenswrapper[4669]: E1010 09:25:38.142268 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e" Oct 10 09:25:38 crc kubenswrapper[4669]: E1010 09:25:38.142409 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8xbn4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-db6d7f97b-hjkht_openstack-operators(db1620d5-baef-433a-b9e7-07a55004a68f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:38 crc kubenswrapper[4669]: E1010 09:25:38.626075 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167" Oct 10 09:25:38 crc kubenswrapper[4669]: E1010 09:25:38.626302 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rzlrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-f9fb45f8f-n72xm_openstack-operators(82353170-7f06-43e4-966b-20d0d234f0f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:39 crc kubenswrapper[4669]: E1010 09:25:39.017978 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960" Oct 10 09:25:39 crc kubenswrapper[4669]: E1010 09:25:39.018159 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w7528,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-9c5c78d49-k5nxp_openstack-operators(2132b635-c43e-4c4d-9074-37e81b6345f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:41 crc kubenswrapper[4669]: E1010 09:25:41.792021 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff" Oct 10 09:25:41 crc kubenswrapper[4669]: E1010 09:25:41.792538 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bdx5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-68b6c87b68-4b4jb_openstack-operators(316d31d2-d9a9-4fd7-a079-4620ad8c5e09): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:42 crc kubenswrapper[4669]: E1010 09:25:42.373112 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0" Oct 10 09:25:42 crc kubenswrapper[4669]: E1010 09:25:42.373517 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hlgsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-858f76bbdd-pdnnc_openstack-operators(bd3dd74a-3c58-4785-854e-400c7ba726de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:42 crc kubenswrapper[4669]: E1010 09:25:42.839881 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8" Oct 10 09:25:42 crc kubenswrapper[4669]: E1010 09:25:42.840094 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4kr5z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5f67fbc655-q52hz_openstack-operators(6941897d-fd2e-4ff3-ad50-aeba82eb86c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:43 crc kubenswrapper[4669]: E1010 09:25:43.274159 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492" Oct 10 09:25:43 crc kubenswrapper[4669]: E1010 09:25:43.274351 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t5fvj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-656bcbd775-czldp_openstack-operators(270b1e97-6178-44d9-8d9f-ed5e4dc02000): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:43 crc kubenswrapper[4669]: E1010 09:25:43.945550 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Oct 10 09:25:43 crc kubenswrapper[4669]: E1010 09:25:43.945833 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g8mbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs_openstack-operators(637efcba-83b3-4f3a-93f0-1ce6257987e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:25:43 crc kubenswrapper[4669]: E1010 09:25:43.947032 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" podUID="637efcba-83b3-4f3a-93f0-1ce6257987e1" Oct 10 09:25:44 crc kubenswrapper[4669]: E1010 09:25:44.554412 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" podUID="637efcba-83b3-4f3a-93f0-1ce6257987e1" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.292610 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" podUID="1a5e08a1-2d18-4382-8129-be1376c40bde" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.473031 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" podUID="27b6cee3-b0a3-4116-b9bb-299a81bba403" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.555543 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" podUID="affd00ea-eac5-4c5f-b452-c8bb0e17aba8" Oct 10 09:25:45 crc kubenswrapper[4669]: I1010 09:25:45.560257 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" event={"ID":"affd00ea-eac5-4c5f-b452-c8bb0e17aba8","Type":"ContainerStarted","Data":"9d4ac08978b71a79ede8770dca06502239588a327a008b6a133721c219b2659c"} Oct 10 09:25:45 crc kubenswrapper[4669]: I1010 09:25:45.562434 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" event={"ID":"27b6cee3-b0a3-4116-b9bb-299a81bba403","Type":"ContainerStarted","Data":"4a2245c6636e176553ef46f94848995c1bbaaa9972684ae5ebe5065d2577024d"} Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.564301 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" podUID="27b6cee3-b0a3-4116-b9bb-299a81bba403" Oct 10 09:25:45 crc kubenswrapper[4669]: I1010 09:25:45.567553 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" event={"ID":"1a5e08a1-2d18-4382-8129-be1376c40bde","Type":"ContainerStarted","Data":"b8f1178dd3c8925892fd0f069ac3d474fced3a51b8296124888275bce8c2d264"} Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.569488 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:73736f216f886549901fbcfc823b072f73691c9a79ec79e59d100e992b9c1e34\\\"\"" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" podUID="1a5e08a1-2d18-4382-8129-be1376c40bde" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.647605 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" podUID="f30a8280-3312-4333-8966-810b8dd3c889" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.647782 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" podUID="bd3dd74a-3c58-4785-854e-400c7ba726de" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.668044 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" podUID="82353170-7f06-43e4-966b-20d0d234f0f2" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.697931 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" podUID="684c8731-c863-49d2-a029-d9d69ae01f8d" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.748015 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" podUID="2132b635-c43e-4c4d-9074-37e81b6345f2" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.801821 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" podUID="43bf605f-c6c8-447b-9dc3-03be42fe7f10" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.899992 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" podUID="db1620d5-baef-433a-b9e7-07a55004a68f" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.913219 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" podUID="316d31d2-d9a9-4fd7-a079-4620ad8c5e09" Oct 10 09:25:45 crc kubenswrapper[4669]: E1010 09:25:45.946050 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" podUID="270b1e97-6178-44d9-8d9f-ed5e4dc02000" Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.182372 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" podUID="6941897d-fd2e-4ff3-ad50-aeba82eb86c2" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.639846 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" event={"ID":"5d2f190a-4eae-44e6-a5cb-e15abf2e037d","Type":"ContainerStarted","Data":"d02438149a35d59968fc4847a5390cb3b7729f2f0528b3e466bb1c205bac4c2e"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.640903 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.679985 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" event={"ID":"684c8731-c863-49d2-a029-d9d69ae01f8d","Type":"ContainerStarted","Data":"9e1e52d790aea4265c31463d1531fa8ae131d1e2ca1c365bc75bc01484c23254"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.695711 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" event={"ID":"bd3dd74a-3c58-4785-854e-400c7ba726de","Type":"ContainerStarted","Data":"e2b97b69a40508d37616df154726a59f37c2ab6d515b18f1f808410e54872f79"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.696097 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" podStartSLOduration=5.053024752 podStartE2EDuration="30.696087721s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.516791786 +0000 UTC m=+862.532810528" lastFinishedPulling="2025-10-10 09:25:45.159854765 +0000 UTC m=+888.175873497" observedRunningTime="2025-10-10 09:25:46.695977848 +0000 UTC m=+889.711996590" watchObservedRunningTime="2025-10-10 09:25:46.696087721 +0000 UTC m=+889.712106463" Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.698547 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" podUID="bd3dd74a-3c58-4785-854e-400c7ba726de" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.707283 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" event={"ID":"db1620d5-baef-433a-b9e7-07a55004a68f","Type":"ContainerStarted","Data":"cd83c8ba1208cafad8d88d8c30adf5ef1829e9d2d477398cd2ff13e00119ad87"} Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.715376 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" podUID="db1620d5-baef-433a-b9e7-07a55004a68f" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.722805 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" event={"ID":"bdebd762-e639-4e50-820d-cd32885a1341","Type":"ContainerStarted","Data":"de6f3e8e733bb4b624841f76f9db5e1d6fe75a2ab05721fbc6bc7097413e0805"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.750785 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" event={"ID":"2132b635-c43e-4c4d-9074-37e81b6345f2","Type":"ContainerStarted","Data":"b9f1fda68fb6d3bb35c68f960c789403de946da1f7dc9b6251f59e4cfb2b2c33"} Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.756200 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" podUID="2132b635-c43e-4c4d-9074-37e81b6345f2" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.766435 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" event={"ID":"29661927-099c-4549-8584-9390881170d5","Type":"ContainerStarted","Data":"c317d616a04eb539b354eeb84f17602ff4deb9abb222e4e02d662f0e3af238a7"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.790846 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" event={"ID":"55f7e993-3dd8-4acf-9716-47cef7f73ef8","Type":"ContainerStarted","Data":"9c1cffc587fee0ae2f5fd1416705bdca4350d6aa762c79156678dce68a779c1d"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.827613 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" event={"ID":"43bf605f-c6c8-447b-9dc3-03be42fe7f10","Type":"ContainerStarted","Data":"bd959790f3ef8fb159e453f75781fff20560f0d805db0b2c1c601b24465c81f9"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.855545 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" event={"ID":"909f202d-6571-4743-9314-26cc27e94c77","Type":"ContainerStarted","Data":"a7f47c2a5b1a957daea0caa28a6b7893a19e841f3c3faf0ed1c6c5eb8cc31bf4"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.883071 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" event={"ID":"316d31d2-d9a9-4fd7-a079-4620ad8c5e09","Type":"ContainerStarted","Data":"3729ef9c2717839efdd1eedbb11434009112cba10d14754017fcd3d3c8019337"} Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.884529 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" podUID="316d31d2-d9a9-4fd7-a079-4620ad8c5e09" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.903774 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" event={"ID":"73105b8d-0527-4510-886b-7203163993fd","Type":"ContainerStarted","Data":"96d56174c291213d8c4142e20032f209fa77cc13e574c89cacb06c2d27f314ed"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.903991 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.907372 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" event={"ID":"f30a8280-3312-4333-8966-810b8dd3c889","Type":"ContainerStarted","Data":"d932a27afc4167524208a21cb978e568675d198aa4b2ce46d5d694c9e5764cfd"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.926874 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" event={"ID":"270b1e97-6178-44d9-8d9f-ed5e4dc02000","Type":"ContainerStarted","Data":"7e320e0d49fb9836a7cd7cbfd174f172ba59de948e7ec340972ad4078ce8a3ad"} Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.936315 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" podUID="270b1e97-6178-44d9-8d9f-ed5e4dc02000" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.947342 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" event={"ID":"6941897d-fd2e-4ff3-ad50-aeba82eb86c2","Type":"ContainerStarted","Data":"c185288e29952cb7a399c137d53b22f6ff9f220f2b48c7cf59642f84f6dd5cd4"} Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.954904 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" podUID="6941897d-fd2e-4ff3-ad50-aeba82eb86c2" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.968769 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" event={"ID":"82353170-7f06-43e4-966b-20d0d234f0f2","Type":"ContainerStarted","Data":"9b48706a83ce0084872385fb7bde2bbb0b439a0a8832cfdde56b3fb3ca00cf1e"} Oct 10 09:25:46 crc kubenswrapper[4669]: E1010 09:25:46.974947 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" podUID="82353170-7f06-43e4-966b-20d0d234f0f2" Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.992810 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" event={"ID":"7dcc64bf-ac2d-4357-adcb-b6c146462464","Type":"ContainerStarted","Data":"f5c8626854930c8ba6e3bba1d8023fb2cd676855a58c08801834304218d99a5b"} Oct 10 09:25:46 crc kubenswrapper[4669]: I1010 09:25:46.993515 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.021089 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" event={"ID":"48aca3a8-684a-448f-ab17-b9d604d47484","Type":"ContainerStarted","Data":"5181c6bee5a3b87cd33e9644909682e58140452bea87cc94e580b602e172388b"} Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.021127 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.021137 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" event={"ID":"48aca3a8-684a-448f-ab17-b9d604d47484","Type":"ContainerStarted","Data":"debf9d089247747700702e7628a0df8b9549d0f51fdf3983e2dd9b408a7b087f"} Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.043546 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fhcl" event={"ID":"178baabb-2525-43c3-b1c3-844f4311f786","Type":"ContainerStarted","Data":"ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa"} Oct 10 09:25:47 crc kubenswrapper[4669]: E1010 09:25:47.049793 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" podUID="27b6cee3-b0a3-4116-b9bb-299a81bba403" Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.110992 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" podStartSLOduration=4.677158328 podStartE2EDuration="31.110974835s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:17.496678994 +0000 UTC m=+860.512697736" lastFinishedPulling="2025-10-10 09:25:43.930495501 +0000 UTC m=+886.946514243" observedRunningTime="2025-10-10 09:25:47.106277432 +0000 UTC m=+890.122296174" watchObservedRunningTime="2025-10-10 09:25:47.110974835 +0000 UTC m=+890.126993567" Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.161835 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" podStartSLOduration=5.628785725 podStartE2EDuration="31.161818508s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.608649451 +0000 UTC m=+862.624668193" lastFinishedPulling="2025-10-10 09:25:45.141682234 +0000 UTC m=+888.157700976" observedRunningTime="2025-10-10 09:25:47.158048545 +0000 UTC m=+890.174067287" watchObservedRunningTime="2025-10-10 09:25:47.161818508 +0000 UTC m=+890.177837250" Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.331109 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" podStartSLOduration=4.858746049 podStartE2EDuration="31.331089678s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.470689449 +0000 UTC m=+861.486708191" lastFinishedPulling="2025-10-10 09:25:44.943033078 +0000 UTC m=+887.959051820" observedRunningTime="2025-10-10 09:25:47.260738562 +0000 UTC m=+890.276757304" watchObservedRunningTime="2025-10-10 09:25:47.331089678 +0000 UTC m=+890.347108420" Oct 10 09:25:47 crc kubenswrapper[4669]: I1010 09:25:47.578387 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8fhcl" podStartSLOduration=4.850996926 podStartE2EDuration="28.578372145s" podCreationTimestamp="2025-10-10 09:25:19 +0000 UTC" firstStartedPulling="2025-10-10 09:25:21.336170505 +0000 UTC m=+864.352189247" lastFinishedPulling="2025-10-10 09:25:45.063545724 +0000 UTC m=+888.079564466" observedRunningTime="2025-10-10 09:25:47.516914878 +0000 UTC m=+890.532933630" watchObservedRunningTime="2025-10-10 09:25:47.578372145 +0000 UTC m=+890.594390887" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.051266 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" event={"ID":"1a5e08a1-2d18-4382-8129-be1376c40bde","Type":"ContainerStarted","Data":"d7488f3015131cf9ea6b2c5fb90524af646e4943f4702a29e73046f22f20787b"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.051501 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.052774 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" event={"ID":"bdebd762-e639-4e50-820d-cd32885a1341","Type":"ContainerStarted","Data":"6b9968f37055bce789144278507c0b20abe8b9821f40ec8d68612f3508d86c5a"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.052893 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.054373 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" event={"ID":"29661927-099c-4549-8584-9390881170d5","Type":"ContainerStarted","Data":"3cc5d52b53dc7eff6183741f3b0a67bcf277e1a636550167b1c98bbcc8eec7df"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.054483 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.056373 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" event={"ID":"55f7e993-3dd8-4acf-9716-47cef7f73ef8","Type":"ContainerStarted","Data":"db4301c0f88e5a4ecbda3d8b2e38311daa39b5b33651fad3c86bc69cbd12e326"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.056788 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.059070 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" event={"ID":"909f202d-6571-4743-9314-26cc27e94c77","Type":"ContainerStarted","Data":"03dcc35e57f177b4f033213e54f55ed0b3bb548eec2f9a34362b2af3b85441c0"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.059457 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.060998 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" event={"ID":"7dcc64bf-ac2d-4357-adcb-b6c146462464","Type":"ContainerStarted","Data":"7db5eb8adbc51f07bc4b546f5a5b068dc139657fcb5accdd2ab7eb25e497aab8"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.063314 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" event={"ID":"affd00ea-eac5-4c5f-b452-c8bb0e17aba8","Type":"ContainerStarted","Data":"59d023f2a93cddee375e67e2791093998db73ffdb0e627ba6161913e9d50c0e9"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.063733 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.065671 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" event={"ID":"f30a8280-3312-4333-8966-810b8dd3c889","Type":"ContainerStarted","Data":"ebd65cdd4a688bbf8d7b5272183d61d232e4b7995dfdadee8e010b476247662f"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.066048 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.067903 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" event={"ID":"43bf605f-c6c8-447b-9dc3-03be42fe7f10","Type":"ContainerStarted","Data":"1e850cee95323cc1c3b84b7a6379054b2273c4382a4779c4450907f847b5bed5"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.068262 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.070779 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" event={"ID":"684c8731-c863-49d2-a029-d9d69ae01f8d","Type":"ContainerStarted","Data":"39ce1d79daa8143c52a04c8b148523e89b5c3ace5c258891576f413bafe0d7fb"} Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.070804 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" Oct 10 09:25:48 crc kubenswrapper[4669]: E1010 09:25:48.074203 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" podUID="270b1e97-6178-44d9-8d9f-ed5e4dc02000" Oct 10 09:25:48 crc kubenswrapper[4669]: E1010 09:25:48.074802 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" podUID="6941897d-fd2e-4ff3-ad50-aeba82eb86c2" Oct 10 09:25:48 crc kubenswrapper[4669]: E1010 09:25:48.074919 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" podUID="bd3dd74a-3c58-4785-854e-400c7ba726de" Oct 10 09:25:48 crc kubenswrapper[4669]: E1010 09:25:48.074954 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" podUID="2132b635-c43e-4c4d-9074-37e81b6345f2" Oct 10 09:25:48 crc kubenswrapper[4669]: E1010 09:25:48.074996 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" podUID="82353170-7f06-43e4-966b-20d0d234f0f2" Oct 10 09:25:48 crc kubenswrapper[4669]: E1010 09:25:48.075227 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" podUID="316d31d2-d9a9-4fd7-a079-4620ad8c5e09" Oct 10 09:25:48 crc kubenswrapper[4669]: E1010 09:25:48.075279 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" podUID="db1620d5-baef-433a-b9e7-07a55004a68f" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.096496 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" podStartSLOduration=2.244102245 podStartE2EDuration="32.096478643s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:17.79195168 +0000 UTC m=+860.807970422" lastFinishedPulling="2025-10-10 09:25:47.644328078 +0000 UTC m=+890.660346820" observedRunningTime="2025-10-10 09:25:48.095145219 +0000 UTC m=+891.111163961" watchObservedRunningTime="2025-10-10 09:25:48.096478643 +0000 UTC m=+891.112497385" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.133889 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" podStartSLOduration=3.471924999 podStartE2EDuration="32.133871798s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.77780302 +0000 UTC m=+861.793821762" lastFinishedPulling="2025-10-10 09:25:47.439749819 +0000 UTC m=+890.455768561" observedRunningTime="2025-10-10 09:25:48.131095638 +0000 UTC m=+891.147114380" watchObservedRunningTime="2025-10-10 09:25:48.133871798 +0000 UTC m=+891.149890540" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.183693 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" podStartSLOduration=4.153325134 podStartE2EDuration="32.183674557s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.245415887 +0000 UTC m=+862.261434619" lastFinishedPulling="2025-10-10 09:25:47.2757653 +0000 UTC m=+890.291784042" observedRunningTime="2025-10-10 09:25:48.178004013 +0000 UTC m=+891.194022755" watchObservedRunningTime="2025-10-10 09:25:48.183674557 +0000 UTC m=+891.199693299" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.204987 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" podStartSLOduration=6.107339208 podStartE2EDuration="32.204970409s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:17.861049606 +0000 UTC m=+860.877068348" lastFinishedPulling="2025-10-10 09:25:43.958680787 +0000 UTC m=+886.974699549" observedRunningTime="2025-10-10 09:25:48.202148737 +0000 UTC m=+891.218167499" watchObservedRunningTime="2025-10-10 09:25:48.204970409 +0000 UTC m=+891.220989151" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.323412 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" podStartSLOduration=6.610327085 podStartE2EDuration="32.323395988s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.725931564 +0000 UTC m=+861.741950306" lastFinishedPulling="2025-10-10 09:25:44.439000467 +0000 UTC m=+887.455019209" observedRunningTime="2025-10-10 09:25:48.322408165 +0000 UTC m=+891.338426907" watchObservedRunningTime="2025-10-10 09:25:48.323395988 +0000 UTC m=+891.339414730" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.365776 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" podStartSLOduration=6.127698991 podStartE2EDuration="32.365753545s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.704034323 +0000 UTC m=+861.720053065" lastFinishedPulling="2025-10-10 09:25:44.942088887 +0000 UTC m=+887.958107619" observedRunningTime="2025-10-10 09:25:48.364802703 +0000 UTC m=+891.380821445" watchObservedRunningTime="2025-10-10 09:25:48.365753545 +0000 UTC m=+891.381772277" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.447245 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" podStartSLOduration=4.147275965 podStartE2EDuration="32.447224371s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:17.996844609 +0000 UTC m=+861.012863351" lastFinishedPulling="2025-10-10 09:25:46.296793015 +0000 UTC m=+889.312811757" observedRunningTime="2025-10-10 09:25:48.405575838 +0000 UTC m=+891.421594580" watchObservedRunningTime="2025-10-10 09:25:48.447224371 +0000 UTC m=+891.463243123" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.476389 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" podStartSLOduration=4.302667148 podStartE2EDuration="32.476372049s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.366698649 +0000 UTC m=+862.382717391" lastFinishedPulling="2025-10-10 09:25:47.54040355 +0000 UTC m=+890.556422292" observedRunningTime="2025-10-10 09:25:48.448269436 +0000 UTC m=+891.464288168" watchObservedRunningTime="2025-10-10 09:25:48.476372049 +0000 UTC m=+891.492390781" Oct 10 09:25:48 crc kubenswrapper[4669]: I1010 09:25:48.502944 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" podStartSLOduration=7.417362063 podStartE2EDuration="32.502929513s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.341271822 +0000 UTC m=+862.357290564" lastFinishedPulling="2025-10-10 09:25:44.426839272 +0000 UTC m=+887.442858014" observedRunningTime="2025-10-10 09:25:48.499398248 +0000 UTC m=+891.515416990" watchObservedRunningTime="2025-10-10 09:25:48.502929513 +0000 UTC m=+891.518948255" Oct 10 09:25:50 crc kubenswrapper[4669]: I1010 09:25:50.083605 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-pfp4n" Oct 10 09:25:50 crc kubenswrapper[4669]: I1010 09:25:50.083929 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-96278" Oct 10 09:25:50 crc kubenswrapper[4669]: I1010 09:25:50.212121 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:50 crc kubenswrapper[4669]: I1010 09:25:50.212298 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:25:51 crc kubenswrapper[4669]: I1010 09:25:51.253300 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8fhcl" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="registry-server" probeResult="failure" output=< Oct 10 09:25:51 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:25:51 crc kubenswrapper[4669]: > Oct 10 09:25:54 crc kubenswrapper[4669]: I1010 09:25:54.275006 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:25:54 crc kubenswrapper[4669]: I1010 09:25:54.275077 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:25:56 crc kubenswrapper[4669]: I1010 09:25:56.451428 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-jzps8" Oct 10 09:25:56 crc kubenswrapper[4669]: I1010 09:25:56.456105 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-pn8xd" Oct 10 09:25:56 crc kubenswrapper[4669]: I1010 09:25:56.713356 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-sbc5v" Oct 10 09:25:56 crc kubenswrapper[4669]: I1010 09:25:56.944135 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-kwn4g" Oct 10 09:25:56 crc kubenswrapper[4669]: I1010 09:25:56.960686 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-pdf8v" Oct 10 09:25:56 crc kubenswrapper[4669]: I1010 09:25:56.984573 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-wtsqj" Oct 10 09:25:57 crc kubenswrapper[4669]: I1010 09:25:57.351132 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-j7fm7" Oct 10 09:25:57 crc kubenswrapper[4669]: I1010 09:25:57.365029 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-r7ctx" Oct 10 09:25:57 crc kubenswrapper[4669]: I1010 09:25:57.482164 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5458f77c4-b2p7l" Oct 10 09:25:57 crc kubenswrapper[4669]: I1010 09:25:57.523112 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-cjtjr" Oct 10 09:25:58 crc kubenswrapper[4669]: I1010 09:25:58.135132 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" event={"ID":"637efcba-83b3-4f3a-93f0-1ce6257987e1","Type":"ContainerStarted","Data":"ee866f72a3822bfe66a94b8b0f55f664a6177811d7befda4744936abb8fe9bd6"} Oct 10 09:25:58 crc kubenswrapper[4669]: I1010 09:25:58.157268 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs" podStartSLOduration=2.83043909 podStartE2EDuration="41.15724326s" podCreationTimestamp="2025-10-10 09:25:17 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.233723287 +0000 UTC m=+862.249742029" lastFinishedPulling="2025-10-10 09:25:57.560527457 +0000 UTC m=+900.576546199" observedRunningTime="2025-10-10 09:25:58.151856135 +0000 UTC m=+901.167874887" watchObservedRunningTime="2025-10-10 09:25:58.15724326 +0000 UTC m=+901.173262012" Oct 10 09:25:58 crc kubenswrapper[4669]: I1010 09:25:58.682262 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv" Oct 10 09:26:00 crc kubenswrapper[4669]: I1010 09:26:00.148990 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" event={"ID":"db1620d5-baef-433a-b9e7-07a55004a68f","Type":"ContainerStarted","Data":"6f80e9d381ad3ebe222f9cb199045ef91d38dd627893354fdb51390b1aa0a44b"} Oct 10 09:26:00 crc kubenswrapper[4669]: I1010 09:26:00.149500 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.155937 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" event={"ID":"270b1e97-6178-44d9-8d9f-ed5e4dc02000","Type":"ContainerStarted","Data":"2a9879ec109043d79471a7f6656b38dfc17aa70c43bc06d2179eedb1d44ef856"} Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.156417 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.157814 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" event={"ID":"bd3dd74a-3c58-4785-854e-400c7ba726de","Type":"ContainerStarted","Data":"99cf2e45cba0f677848b58e339da6b04f738abc48a53b5e180992877e1da98e6"} Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.158079 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.159387 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" event={"ID":"6941897d-fd2e-4ff3-ad50-aeba82eb86c2","Type":"ContainerStarted","Data":"d12cf8989efae441b79223ccf3eaf3128e8c873dbb0c90a39cdcd54e53720af2"} Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.159649 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.182842 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" podStartSLOduration=5.48433762 podStartE2EDuration="45.182823759s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.607859296 +0000 UTC m=+862.623878038" lastFinishedPulling="2025-10-10 09:25:59.306345425 +0000 UTC m=+902.322364177" observedRunningTime="2025-10-10 09:26:00.197344352 +0000 UTC m=+903.213363094" watchObservedRunningTime="2025-10-10 09:26:01.182823759 +0000 UTC m=+904.198842501" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.184435 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" podStartSLOduration=4.40244378 podStartE2EDuration="45.184426042s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.424037922 +0000 UTC m=+862.440056664" lastFinishedPulling="2025-10-10 09:26:00.206020184 +0000 UTC m=+903.222038926" observedRunningTime="2025-10-10 09:26:01.176198104 +0000 UTC m=+904.192216856" watchObservedRunningTime="2025-10-10 09:26:01.184426042 +0000 UTC m=+904.200444784" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.200844 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" podStartSLOduration=2.974187563 podStartE2EDuration="45.200826505s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.190625457 +0000 UTC m=+861.206644199" lastFinishedPulling="2025-10-10 09:26:00.417264399 +0000 UTC m=+903.433283141" observedRunningTime="2025-10-10 09:26:01.196222525 +0000 UTC m=+904.212241277" watchObservedRunningTime="2025-10-10 09:26:01.200826505 +0000 UTC m=+904.216845247" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.217781 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" podStartSLOduration=3.732417394 podStartE2EDuration="45.217762785s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.698988318 +0000 UTC m=+861.715007060" lastFinishedPulling="2025-10-10 09:26:00.184333709 +0000 UTC m=+903.200352451" observedRunningTime="2025-10-10 09:26:01.212389051 +0000 UTC m=+904.228407793" watchObservedRunningTime="2025-10-10 09:26:01.217762785 +0000 UTC m=+904.233781527" Oct 10 09:26:01 crc kubenswrapper[4669]: I1010 09:26:01.347797 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8fhcl" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="registry-server" probeResult="failure" output=< Oct 10 09:26:01 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:26:01 crc kubenswrapper[4669]: > Oct 10 09:26:02 crc kubenswrapper[4669]: I1010 09:26:02.176059 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" event={"ID":"2132b635-c43e-4c4d-9074-37e81b6345f2","Type":"ContainerStarted","Data":"02fcd625707e97b812c971836336146b6f945ee3845b8c68ab279bf573016fa6"} Oct 10 09:26:02 crc kubenswrapper[4669]: I1010 09:26:02.176717 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" Oct 10 09:26:02 crc kubenswrapper[4669]: I1010 09:26:02.178320 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" event={"ID":"82353170-7f06-43e4-966b-20d0d234f0f2","Type":"ContainerStarted","Data":"896429457c3dc99c961f4208192e4e2e25bbd5b1760b199b3802b40888b4c166"} Oct 10 09:26:02 crc kubenswrapper[4669]: I1010 09:26:02.178750 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" Oct 10 09:26:02 crc kubenswrapper[4669]: I1010 09:26:02.218743 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" podStartSLOduration=3.648515117 podStartE2EDuration="46.218728835s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.798685509 +0000 UTC m=+861.814704251" lastFinishedPulling="2025-10-10 09:26:01.368899227 +0000 UTC m=+904.384917969" observedRunningTime="2025-10-10 09:26:02.198949472 +0000 UTC m=+905.214968224" watchObservedRunningTime="2025-10-10 09:26:02.218728835 +0000 UTC m=+905.234747577" Oct 10 09:26:02 crc kubenswrapper[4669]: I1010 09:26:02.222154 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" podStartSLOduration=3.517311802 podStartE2EDuration="46.222140086s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.663114282 +0000 UTC m=+861.679133024" lastFinishedPulling="2025-10-10 09:26:01.367942556 +0000 UTC m=+904.383961308" observedRunningTime="2025-10-10 09:26:02.219825471 +0000 UTC m=+905.235844223" watchObservedRunningTime="2025-10-10 09:26:02.222140086 +0000 UTC m=+905.238158828" Oct 10 09:26:03 crc kubenswrapper[4669]: I1010 09:26:03.185739 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" event={"ID":"27b6cee3-b0a3-4116-b9bb-299a81bba403","Type":"ContainerStarted","Data":"e136646a6a8d811abb4b2bbc265debd4cc64236dca2aff594079df53cc61d9da"} Oct 10 09:26:03 crc kubenswrapper[4669]: I1010 09:26:03.186255 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" Oct 10 09:26:05 crc kubenswrapper[4669]: I1010 09:26:05.203175 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" event={"ID":"316d31d2-d9a9-4fd7-a079-4620ad8c5e09","Type":"ContainerStarted","Data":"5927c0c602cf22b895b9bf1670c96818f2f884818335f5ea37152fbd3ade48df"} Oct 10 09:26:05 crc kubenswrapper[4669]: I1010 09:26:05.204276 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" Oct 10 09:26:05 crc kubenswrapper[4669]: I1010 09:26:05.218266 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" podStartSLOduration=4.15903528 podStartE2EDuration="49.218246782s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:19.480722664 +0000 UTC m=+862.496741406" lastFinishedPulling="2025-10-10 09:26:04.539934166 +0000 UTC m=+907.555952908" observedRunningTime="2025-10-10 09:26:05.218195871 +0000 UTC m=+908.234214623" watchObservedRunningTime="2025-10-10 09:26:05.218246782 +0000 UTC m=+908.234265534" Oct 10 09:26:05 crc kubenswrapper[4669]: I1010 09:26:05.223505 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" podStartSLOduration=5.376387626 podStartE2EDuration="49.223466882s" podCreationTimestamp="2025-10-10 09:25:16 +0000 UTC" firstStartedPulling="2025-10-10 09:25:18.611223746 +0000 UTC m=+861.627242488" lastFinishedPulling="2025-10-10 09:26:02.458303002 +0000 UTC m=+905.474321744" observedRunningTime="2025-10-10 09:26:03.204602275 +0000 UTC m=+906.220621027" watchObservedRunningTime="2025-10-10 09:26:05.223466882 +0000 UTC m=+908.239485644" Oct 10 09:26:06 crc kubenswrapper[4669]: I1010 09:26:06.654162 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-pdnnc" Oct 10 09:26:06 crc kubenswrapper[4669]: I1010 09:26:06.809406 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-k5nxp" Oct 10 09:26:06 crc kubenswrapper[4669]: I1010 09:26:06.905892 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-q52hz" Oct 10 09:26:06 crc kubenswrapper[4669]: I1010 09:26:06.916084 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-n72xm" Oct 10 09:26:07 crc kubenswrapper[4669]: I1010 09:26:07.115085 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-hjkht" Oct 10 09:26:07 crc kubenswrapper[4669]: I1010 09:26:07.290188 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-czldp" Oct 10 09:26:11 crc kubenswrapper[4669]: I1010 09:26:11.250879 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8fhcl" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="registry-server" probeResult="failure" output=< Oct 10 09:26:11 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:26:11 crc kubenswrapper[4669]: > Oct 10 09:26:16 crc kubenswrapper[4669]: I1010 09:26:16.853637 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-6zprx" Oct 10 09:26:17 crc kubenswrapper[4669]: I1010 09:26:17.263028 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-4b4jb" Oct 10 09:26:20 crc kubenswrapper[4669]: I1010 09:26:20.279020 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:26:20 crc kubenswrapper[4669]: I1010 09:26:20.335096 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:26:24 crc kubenswrapper[4669]: I1010 09:26:24.275010 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:26:24 crc kubenswrapper[4669]: I1010 09:26:24.275805 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.235843 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8fhcl"] Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.236418 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8fhcl" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="registry-server" containerID="cri-o://ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa" gracePeriod=2 Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.616013 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.730966 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-utilities\") pod \"178baabb-2525-43c3-b1c3-844f4311f786\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.731062 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nvxr\" (UniqueName: \"kubernetes.io/projected/178baabb-2525-43c3-b1c3-844f4311f786-kube-api-access-4nvxr\") pod \"178baabb-2525-43c3-b1c3-844f4311f786\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.731142 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-catalog-content\") pod \"178baabb-2525-43c3-b1c3-844f4311f786\" (UID: \"178baabb-2525-43c3-b1c3-844f4311f786\") " Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.732054 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-utilities" (OuterVolumeSpecName: "utilities") pod "178baabb-2525-43c3-b1c3-844f4311f786" (UID: "178baabb-2525-43c3-b1c3-844f4311f786"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.738416 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178baabb-2525-43c3-b1c3-844f4311f786-kube-api-access-4nvxr" (OuterVolumeSpecName: "kube-api-access-4nvxr") pod "178baabb-2525-43c3-b1c3-844f4311f786" (UID: "178baabb-2525-43c3-b1c3-844f4311f786"). InnerVolumeSpecName "kube-api-access-4nvxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.833025 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nvxr\" (UniqueName: \"kubernetes.io/projected/178baabb-2525-43c3-b1c3-844f4311f786-kube-api-access-4nvxr\") on node \"crc\" DevicePath \"\"" Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.833057 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.846922 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "178baabb-2525-43c3-b1c3-844f4311f786" (UID: "178baabb-2525-43c3-b1c3-844f4311f786"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:26:27 crc kubenswrapper[4669]: I1010 09:26:27.935096 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/178baabb-2525-43c3-b1c3-844f4311f786-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.359915 4669 generic.go:334] "Generic (PLEG): container finished" podID="178baabb-2525-43c3-b1c3-844f4311f786" containerID="ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa" exitCode=0 Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.359955 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fhcl" event={"ID":"178baabb-2525-43c3-b1c3-844f4311f786","Type":"ContainerDied","Data":"ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa"} Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.359982 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fhcl" event={"ID":"178baabb-2525-43c3-b1c3-844f4311f786","Type":"ContainerDied","Data":"18d98314612b7b3d7ed94a7edbfef7897d25367bc1a6e90a571bdcf73f5260cf"} Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.359999 4669 scope.go:117] "RemoveContainer" containerID="ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.360110 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fhcl" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.383886 4669 scope.go:117] "RemoveContainer" containerID="97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.395981 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8fhcl"] Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.401826 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8fhcl"] Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.410787 4669 scope.go:117] "RemoveContainer" containerID="3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.430633 4669 scope.go:117] "RemoveContainer" containerID="ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa" Oct 10 09:26:28 crc kubenswrapper[4669]: E1010 09:26:28.431069 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa\": container with ID starting with ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa not found: ID does not exist" containerID="ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.431108 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa"} err="failed to get container status \"ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa\": rpc error: code = NotFound desc = could not find container \"ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa\": container with ID starting with ad18a827f50850e4d7f592c86e4a835141e7cc1794fda63170ceeabf544bf3aa not found: ID does not exist" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.431134 4669 scope.go:117] "RemoveContainer" containerID="97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117" Oct 10 09:26:28 crc kubenswrapper[4669]: E1010 09:26:28.431452 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117\": container with ID starting with 97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117 not found: ID does not exist" containerID="97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.431481 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117"} err="failed to get container status \"97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117\": rpc error: code = NotFound desc = could not find container \"97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117\": container with ID starting with 97cfa70040a7c0aedf60910b9e531e3530a20b544dfeb5ad1b3f76eafd9a8117 not found: ID does not exist" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.431499 4669 scope.go:117] "RemoveContainer" containerID="3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517" Oct 10 09:26:28 crc kubenswrapper[4669]: E1010 09:26:28.432089 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517\": container with ID starting with 3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517 not found: ID does not exist" containerID="3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517" Oct 10 09:26:28 crc kubenswrapper[4669]: I1010 09:26:28.432150 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517"} err="failed to get container status \"3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517\": rpc error: code = NotFound desc = could not find container \"3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517\": container with ID starting with 3caf0f8bfad9da8b5765e916d1d9163494fb282a5c37b066b5fccfec66181517 not found: ID does not exist" Oct 10 09:26:29 crc kubenswrapper[4669]: I1010 09:26:29.805904 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="178baabb-2525-43c3-b1c3-844f4311f786" path="/var/lib/kubelet/pods/178baabb-2525-43c3-b1c3-844f4311f786/volumes" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.154175 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-6xz9c"] Oct 10 09:26:40 crc kubenswrapper[4669]: E1010 09:26:40.155792 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="extract-content" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.155863 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="extract-content" Oct 10 09:26:40 crc kubenswrapper[4669]: E1010 09:26:40.155935 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="registry-server" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.155989 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="registry-server" Oct 10 09:26:40 crc kubenswrapper[4669]: E1010 09:26:40.156062 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="registry-server" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.156134 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="registry-server" Oct 10 09:26:40 crc kubenswrapper[4669]: E1010 09:26:40.156195 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="extract-content" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.156251 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="extract-content" Oct 10 09:26:40 crc kubenswrapper[4669]: E1010 09:26:40.156312 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="extract-utilities" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.156362 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="extract-utilities" Oct 10 09:26:40 crc kubenswrapper[4669]: E1010 09:26:40.156421 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="extract-utilities" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.156474 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="extract-utilities" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.156674 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="178baabb-2525-43c3-b1c3-844f4311f786" containerName="registry-server" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.156740 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c07ede9-cc73-43c5-812b-0db964c66349" containerName="registry-server" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.157629 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.161755 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-nng6d" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.161790 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.161834 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.163348 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.167526 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-6xz9c"] Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.241102 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qsllq"] Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.242210 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.243999 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.260974 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qsllq"] Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.301486 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-config\") pod \"dnsmasq-dns-7bfcb9d745-6xz9c\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.301535 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.301763 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxqbv\" (UniqueName: \"kubernetes.io/projected/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-kube-api-access-rxqbv\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.301812 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-config\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.301853 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57fz7\" (UniqueName: \"kubernetes.io/projected/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-kube-api-access-57fz7\") pod \"dnsmasq-dns-7bfcb9d745-6xz9c\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.403470 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxqbv\" (UniqueName: \"kubernetes.io/projected/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-kube-api-access-rxqbv\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.403533 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-config\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.403570 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57fz7\" (UniqueName: \"kubernetes.io/projected/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-kube-api-access-57fz7\") pod \"dnsmasq-dns-7bfcb9d745-6xz9c\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.403628 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-config\") pod \"dnsmasq-dns-7bfcb9d745-6xz9c\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.403651 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.404719 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-dns-svc\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.404789 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-config\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.404789 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-config\") pod \"dnsmasq-dns-7bfcb9d745-6xz9c\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.426555 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57fz7\" (UniqueName: \"kubernetes.io/projected/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-kube-api-access-57fz7\") pod \"dnsmasq-dns-7bfcb9d745-6xz9c\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.426636 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxqbv\" (UniqueName: \"kubernetes.io/projected/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-kube-api-access-rxqbv\") pod \"dnsmasq-dns-758b79db4c-qsllq\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.478340 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.559792 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.904021 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qsllq"] Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.952508 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-f75sk"] Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.953672 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.966424 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-6xz9c"] Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.975362 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:26:40 crc kubenswrapper[4669]: I1010 09:26:40.983060 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-f75sk"] Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.019451 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-config\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.019515 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7d8k\" (UniqueName: \"kubernetes.io/projected/739d7b10-7e0a-4d33-9be2-2c6676088dbc-kube-api-access-j7d8k\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.019559 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-dns-svc\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.046003 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qsllq"] Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.121286 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-config\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.121338 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7d8k\" (UniqueName: \"kubernetes.io/projected/739d7b10-7e0a-4d33-9be2-2c6676088dbc-kube-api-access-j7d8k\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.121367 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-dns-svc\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.122150 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-dns-svc\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.122676 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-config\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.148323 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7d8k\" (UniqueName: \"kubernetes.io/projected/739d7b10-7e0a-4d33-9be2-2c6676088dbc-kube-api-access-j7d8k\") pod \"dnsmasq-dns-8575fc99d7-f75sk\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.284354 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.468097 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-qsllq" event={"ID":"1cccc3a1-eac2-4afe-b25c-27d03564ba8d","Type":"ContainerStarted","Data":"493937a410540711d6a7bc825978e109521008bb2a86ec8080e31906b47b5de1"} Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.469311 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" event={"ID":"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5","Type":"ContainerStarted","Data":"b369b9773bfa36be506c34ee10a84ac50555ebb296212f7dd34db2994d26284f"} Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.477613 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-6xz9c"] Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.575218 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-6jmnh"] Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.576577 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.614958 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-6jmnh"] Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.746469 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78vkp\" (UniqueName: \"kubernetes.io/projected/eafc1e37-eb15-4615-b155-1eb20365e0b5-kube-api-access-78vkp\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.746559 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-dns-svc\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.746654 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-config\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.848135 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-config\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.848221 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78vkp\" (UniqueName: \"kubernetes.io/projected/eafc1e37-eb15-4615-b155-1eb20365e0b5-kube-api-access-78vkp\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.848239 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-dns-svc\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.849149 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-config\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.849167 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-dns-svc\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.867655 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78vkp\" (UniqueName: \"kubernetes.io/projected/eafc1e37-eb15-4615-b155-1eb20365e0b5-kube-api-access-78vkp\") pod \"dnsmasq-dns-77597f887-6jmnh\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:41 crc kubenswrapper[4669]: I1010 09:26:41.896887 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.005929 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-f75sk"] Oct 10 09:26:42 crc kubenswrapper[4669]: W1010 09:26:42.020997 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739d7b10_7e0a_4d33_9be2_2c6676088dbc.slice/crio-9226a2a596aa547759e7c97637d0c59259eab9bdf03653a51381e055e73cd494 WatchSource:0}: Error finding container 9226a2a596aa547759e7c97637d0c59259eab9bdf03653a51381e055e73cd494: Status 404 returned error can't find the container with id 9226a2a596aa547759e7c97637d0c59259eab9bdf03653a51381e055e73cd494 Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.150615 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.159707 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.162707 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.167950 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.168371 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-9rdrl" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.168663 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.168848 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.168890 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.168858 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.181533 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261332 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261371 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c46f84af-548a-4390-8242-c48ff4fc25aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261390 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261555 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c46f84af-548a-4390-8242-c48ff4fc25aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261654 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261691 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261704 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261745 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2rhr\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-kube-api-access-c2rhr\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261765 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261782 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.261804 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363019 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363084 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363103 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363120 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c46f84af-548a-4390-8242-c48ff4fc25aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363374 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c46f84af-548a-4390-8242-c48ff4fc25aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363401 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363423 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363458 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363546 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2rhr\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-kube-api-access-c2rhr\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363562 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363577 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.363942 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.367475 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.368278 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.370517 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.373515 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.374035 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c46f84af-548a-4390-8242-c48ff4fc25aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.378793 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-6jmnh"] Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.380540 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.380653 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c46f84af-548a-4390-8242-c48ff4fc25aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.381164 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.382014 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.389820 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2rhr\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-kube-api-access-c2rhr\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.417551 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.502156 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" event={"ID":"739d7b10-7e0a-4d33-9be2-2c6676088dbc","Type":"ContainerStarted","Data":"9226a2a596aa547759e7c97637d0c59259eab9bdf03653a51381e055e73cd494"} Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.504460 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-6jmnh" event={"ID":"eafc1e37-eb15-4615-b155-1eb20365e0b5","Type":"ContainerStarted","Data":"c7449585cfbf9681d7d362b3d4b5dc6ceef2f5acd32d6a888e6c319d7ee8b52f"} Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.511082 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.701066 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.702351 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.706200 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.707114 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.707818 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-drd74" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.707918 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.708002 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.708117 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.708191 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.721137 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777741 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777778 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777819 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2m77\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-kube-api-access-j2m77\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777841 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777865 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b787ed5f-cbcf-4144-a87f-a014dbfd5282-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777909 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-config-data\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777925 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b787ed5f-cbcf-4144-a87f-a014dbfd5282-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.777972 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.778004 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.778021 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.778052 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.879872 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.879929 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.879950 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.879968 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2m77\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-kube-api-access-j2m77\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.879989 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.880012 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b787ed5f-cbcf-4144-a87f-a014dbfd5282-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.880042 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-config-data\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.880058 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b787ed5f-cbcf-4144-a87f-a014dbfd5282-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.880087 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.880118 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.880135 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.880432 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.881190 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.881344 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.884288 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.885721 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.892464 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.893224 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-config-data\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.903255 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2m77\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-kube-api-access-j2m77\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.905323 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b787ed5f-cbcf-4144-a87f-a014dbfd5282-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.907979 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b787ed5f-cbcf-4144-a87f-a014dbfd5282-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.914646 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:42 crc kubenswrapper[4669]: I1010 09:26:42.944473 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " pod="openstack/rabbitmq-server-0" Oct 10 09:26:43 crc kubenswrapper[4669]: I1010 09:26:43.031528 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 09:26:43 crc kubenswrapper[4669]: I1010 09:26:43.149364 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:26:43 crc kubenswrapper[4669]: I1010 09:26:43.523670 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:26:43 crc kubenswrapper[4669]: W1010 09:26:43.556695 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb787ed5f_cbcf_4144_a87f_a014dbfd5282.slice/crio-5309828c5f2cfa424e41988a4bf0ee40352f6d3e358be206e07bb604e0a1e1a0 WatchSource:0}: Error finding container 5309828c5f2cfa424e41988a4bf0ee40352f6d3e358be206e07bb604e0a1e1a0: Status 404 returned error can't find the container with id 5309828c5f2cfa424e41988a4bf0ee40352f6d3e358be206e07bb604e0a1e1a0 Oct 10 09:26:43 crc kubenswrapper[4669]: I1010 09:26:43.569118 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c46f84af-548a-4390-8242-c48ff4fc25aa","Type":"ContainerStarted","Data":"8844181a1c7977635a20bf8cbb536885fe0b9c3fe8d3afc8b91e19cbfb50fc07"} Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.151111 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.157214 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.162679 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.162932 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.163292 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mcwzc" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.163455 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.163623 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.164892 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.174214 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.211254 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216089 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216172 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216194 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216217 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-secrets\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216248 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216288 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216314 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrjf9\" (UniqueName: \"kubernetes.io/projected/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-kube-api-access-vrjf9\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.216346 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.317467 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.317515 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.317553 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.318243 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.318280 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-secrets\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.318307 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.318327 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.318347 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrjf9\" (UniqueName: \"kubernetes.io/projected/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-kube-api-access-vrjf9\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.318372 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.318973 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.319391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.319096 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.320262 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.322270 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.326226 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.336750 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-secrets\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.346261 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrjf9\" (UniqueName: \"kubernetes.io/projected/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-kube-api-access-vrjf9\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.346504 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79d402e9-8187-4cd5-9df5-9297d8fc4b5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.360426 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"79d402e9-8187-4cd5-9df5-9297d8fc4b5a\") " pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.515508 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 09:26:44 crc kubenswrapper[4669]: I1010 09:26:44.621719 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b787ed5f-cbcf-4144-a87f-a014dbfd5282","Type":"ContainerStarted","Data":"5309828c5f2cfa424e41988a4bf0ee40352f6d3e358be206e07bb604e0a1e1a0"} Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.183397 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 09:26:45 crc kubenswrapper[4669]: W1010 09:26:45.199381 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79d402e9_8187_4cd5_9df5_9297d8fc4b5a.slice/crio-84c83ecdb60897a3d5ef7f96e33604a05a968174708154e4d93419bb738d4cbf WatchSource:0}: Error finding container 84c83ecdb60897a3d5ef7f96e33604a05a968174708154e4d93419bb738d4cbf: Status 404 returned error can't find the container with id 84c83ecdb60897a3d5ef7f96e33604a05a968174708154e4d93419bb738d4cbf Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.503019 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.510949 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.514153 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-7g66d" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.514444 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.514675 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.514890 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.521018 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654374 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02229a1f-8349-4e0c-b5b5-b56ba246f395-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654429 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654456 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654472 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654511 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654554 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n67g2\" (UniqueName: \"kubernetes.io/projected/02229a1f-8349-4e0c-b5b5-b56ba246f395-kube-api-access-n67g2\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654603 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654622 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.654644 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.677792 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"79d402e9-8187-4cd5-9df5-9297d8fc4b5a","Type":"ContainerStarted","Data":"84c83ecdb60897a3d5ef7f96e33604a05a968174708154e4d93419bb738d4cbf"} Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.756725 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n67g2\" (UniqueName: \"kubernetes.io/projected/02229a1f-8349-4e0c-b5b5-b56ba246f395-kube-api-access-n67g2\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.756787 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.756988 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.757007 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.757031 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02229a1f-8349-4e0c-b5b5-b56ba246f395-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.757056 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.757073 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.757090 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.757126 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.758243 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.763517 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02229a1f-8349-4e0c-b5b5-b56ba246f395-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.768376 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.778391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.781482 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.782482 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.784766 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02229a1f-8349-4e0c-b5b5-b56ba246f395-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.802141 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n67g2\" (UniqueName: \"kubernetes.io/projected/02229a1f-8349-4e0c-b5b5-b56ba246f395-kube-api-access-n67g2\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.803622 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02229a1f-8349-4e0c-b5b5-b56ba246f395-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.829610 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"02229a1f-8349-4e0c-b5b5-b56ba246f395\") " pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.842694 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.855693 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.856962 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.862619 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.862982 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-72rk7" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.874037 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.871230 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.960309 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grhfs\" (UniqueName: \"kubernetes.io/projected/9c376674-4cf1-40c0-9e3c-c25e3fefad30-kube-api-access-grhfs\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.960653 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c376674-4cf1-40c0-9e3c-c25e3fefad30-config-data\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.960710 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c376674-4cf1-40c0-9e3c-c25e3fefad30-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.960739 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c376674-4cf1-40c0-9e3c-c25e3fefad30-kolla-config\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:45 crc kubenswrapper[4669]: I1010 09:26:45.960769 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c376674-4cf1-40c0-9e3c-c25e3fefad30-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.062292 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c376674-4cf1-40c0-9e3c-c25e3fefad30-config-data\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.062356 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c376674-4cf1-40c0-9e3c-c25e3fefad30-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.062395 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c376674-4cf1-40c0-9e3c-c25e3fefad30-kolla-config\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.062423 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c376674-4cf1-40c0-9e3c-c25e3fefad30-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.062468 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grhfs\" (UniqueName: \"kubernetes.io/projected/9c376674-4cf1-40c0-9e3c-c25e3fefad30-kube-api-access-grhfs\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.066028 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9c376674-4cf1-40c0-9e3c-c25e3fefad30-kolla-config\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.068263 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c376674-4cf1-40c0-9e3c-c25e3fefad30-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.069340 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c376674-4cf1-40c0-9e3c-c25e3fefad30-config-data\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.079982 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c376674-4cf1-40c0-9e3c-c25e3fefad30-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.081159 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grhfs\" (UniqueName: \"kubernetes.io/projected/9c376674-4cf1-40c0-9e3c-c25e3fefad30-kube-api-access-grhfs\") pod \"memcached-0\" (UID: \"9c376674-4cf1-40c0-9e3c-c25e3fefad30\") " pod="openstack/memcached-0" Oct 10 09:26:46 crc kubenswrapper[4669]: I1010 09:26:46.222607 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.104771 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.288020 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.449435 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.450368 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.453943 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-c5fb5" Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.491437 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.491445 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l9k8\" (UniqueName: \"kubernetes.io/projected/e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79-kube-api-access-2l9k8\") pod \"kube-state-metrics-0\" (UID: \"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79\") " pod="openstack/kube-state-metrics-0" Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.593427 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l9k8\" (UniqueName: \"kubernetes.io/projected/e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79-kube-api-access-2l9k8\") pod \"kube-state-metrics-0\" (UID: \"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79\") " pod="openstack/kube-state-metrics-0" Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.662757 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l9k8\" (UniqueName: \"kubernetes.io/projected/e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79-kube-api-access-2l9k8\") pod \"kube-state-metrics-0\" (UID: \"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79\") " pod="openstack/kube-state-metrics-0" Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.744393 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9c376674-4cf1-40c0-9e3c-c25e3fefad30","Type":"ContainerStarted","Data":"8b779d9686e68d1e31a25ed84ea014662001d25b3f98329f9e000ebd6106b119"} Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.763786 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"02229a1f-8349-4e0c-b5b5-b56ba246f395","Type":"ContainerStarted","Data":"4b291dfeb4d4a6873cb0a06cb2ef358ef7f27d67f241407b294919fa8c90e95e"} Oct 10 09:26:47 crc kubenswrapper[4669]: I1010 09:26:47.791348 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 09:26:48 crc kubenswrapper[4669]: I1010 09:26:48.457132 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:26:48 crc kubenswrapper[4669]: W1010 09:26:48.532688 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8508ed5_1c27_4ce7_b5da_abfbb3bb0b79.slice/crio-84363c2a9cb53453a784ec043bfe54489c4a63d2070cbf73761f8c2bbf68f56f WatchSource:0}: Error finding container 84363c2a9cb53453a784ec043bfe54489c4a63d2070cbf73761f8c2bbf68f56f: Status 404 returned error can't find the container with id 84363c2a9cb53453a784ec043bfe54489c4a63d2070cbf73761f8c2bbf68f56f Oct 10 09:26:48 crc kubenswrapper[4669]: I1010 09:26:48.775073 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79","Type":"ContainerStarted","Data":"84363c2a9cb53453a784ec043bfe54489c4a63d2070cbf73761f8c2bbf68f56f"} Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.296731 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jrw5q"] Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.298434 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.301615 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.301927 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bfpxv" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.303058 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.322245 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-d8wg9"] Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.324195 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.340418 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jrw5q"] Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.345798 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-d8wg9"] Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.372783 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-run\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.372843 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4tt\" (UniqueName: \"kubernetes.io/projected/2b815c31-96c7-47c3-b458-35945b55b326-kube-api-access-km4tt\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.372879 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b815c31-96c7-47c3-b458-35945b55b326-combined-ca-bundle\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.372928 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-run-ovn\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.372970 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b815c31-96c7-47c3-b458-35945b55b326-scripts\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.373005 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b815c31-96c7-47c3-b458-35945b55b326-ovn-controller-tls-certs\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.373023 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-log-ovn\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474701 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-etc-ovs\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474745 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-run-ovn\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474782 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-log\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474800 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b815c31-96c7-47c3-b458-35945b55b326-scripts\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474817 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-lib\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474846 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b815c31-96c7-47c3-b458-35945b55b326-ovn-controller-tls-certs\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474864 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-log-ovn\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474882 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db075709-6fea-46c4-b711-c5517eaae350-scripts\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474932 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-run\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474947 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4tt\" (UniqueName: \"kubernetes.io/projected/2b815c31-96c7-47c3-b458-35945b55b326-kube-api-access-km4tt\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474969 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pshvn\" (UniqueName: \"kubernetes.io/projected/db075709-6fea-46c4-b711-c5517eaae350-kube-api-access-pshvn\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.474987 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b815c31-96c7-47c3-b458-35945b55b326-combined-ca-bundle\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.475007 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-run\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.475302 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-run-ovn\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.475855 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-log-ovn\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.476067 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2b815c31-96c7-47c3-b458-35945b55b326-var-run\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.476965 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b815c31-96c7-47c3-b458-35945b55b326-scripts\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.480578 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b815c31-96c7-47c3-b458-35945b55b326-ovn-controller-tls-certs\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.494605 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b815c31-96c7-47c3-b458-35945b55b326-combined-ca-bundle\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.512258 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4tt\" (UniqueName: \"kubernetes.io/projected/2b815c31-96c7-47c3-b458-35945b55b326-kube-api-access-km4tt\") pod \"ovn-controller-jrw5q\" (UID: \"2b815c31-96c7-47c3-b458-35945b55b326\") " pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.575833 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pshvn\" (UniqueName: \"kubernetes.io/projected/db075709-6fea-46c4-b711-c5517eaae350-kube-api-access-pshvn\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.575943 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-run\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.575986 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-etc-ovs\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.576017 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-log\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.576034 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-lib\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.576059 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db075709-6fea-46c4-b711-c5517eaae350-scripts\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.576207 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-run\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.576358 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-log\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.576400 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-var-lib\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.577006 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/db075709-6fea-46c4-b711-c5517eaae350-etc-ovs\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.577987 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db075709-6fea-46c4-b711-c5517eaae350-scripts\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.629373 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pshvn\" (UniqueName: \"kubernetes.io/projected/db075709-6fea-46c4-b711-c5517eaae350-kube-api-access-pshvn\") pod \"ovn-controller-ovs-d8wg9\" (UID: \"db075709-6fea-46c4-b711-c5517eaae350\") " pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.633444 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q" Oct 10 09:26:50 crc kubenswrapper[4669]: I1010 09:26:50.642723 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.456038 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.458407 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.460747 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-555f6" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.461901 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.462735 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.462883 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.463949 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.471028 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531455 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531522 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531553 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531570 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531614 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5321fb79-5d79-4094-b314-8914ed41cfd5-config\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531640 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjrlh\" (UniqueName: \"kubernetes.io/projected/5321fb79-5d79-4094-b314-8914ed41cfd5-kube-api-access-vjrlh\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531664 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5321fb79-5d79-4094-b314-8914ed41cfd5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.531682 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5321fb79-5d79-4094-b314-8914ed41cfd5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632651 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632703 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632722 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632747 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5321fb79-5d79-4094-b314-8914ed41cfd5-config\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632780 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjrlh\" (UniqueName: \"kubernetes.io/projected/5321fb79-5d79-4094-b314-8914ed41cfd5-kube-api-access-vjrlh\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632812 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5321fb79-5d79-4094-b314-8914ed41cfd5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632838 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5321fb79-5d79-4094-b314-8914ed41cfd5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.632889 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.633290 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.635048 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5321fb79-5d79-4094-b314-8914ed41cfd5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.635961 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5321fb79-5d79-4094-b314-8914ed41cfd5-config\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.636986 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5321fb79-5d79-4094-b314-8914ed41cfd5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.640456 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.650211 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.651171 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjrlh\" (UniqueName: \"kubernetes.io/projected/5321fb79-5d79-4094-b314-8914ed41cfd5-kube-api-access-vjrlh\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.660793 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.669043 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5321fb79-5d79-4094-b314-8914ed41cfd5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5321fb79-5d79-4094-b314-8914ed41cfd5\") " pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:52 crc kubenswrapper[4669]: I1010 09:26:52.787124 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.274625 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.274965 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.275008 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.275718 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"460c2cb0159213ee8a9812e151d381634c493f7eb4c849db7277b1ec7a3b7419"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.275789 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://460c2cb0159213ee8a9812e151d381634c493f7eb4c849db7277b1ec7a3b7419" gracePeriod=600 Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.863107 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="460c2cb0159213ee8a9812e151d381634c493f7eb4c849db7277b1ec7a3b7419" exitCode=0 Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.863154 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"460c2cb0159213ee8a9812e151d381634c493f7eb4c849db7277b1ec7a3b7419"} Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.863190 4669 scope.go:117] "RemoveContainer" containerID="b1be725249e1775acac8f350a1439ffc4d89a22bc933af6755f8ceefcd7eb018" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.906526 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.908563 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.912248 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-d58b4" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.914911 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.915170 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.915197 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.919119 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.968416 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.968499 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f30537-9153-4862-97aa-31e4f5e57ecf-config\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.968641 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7wwn\" (UniqueName: \"kubernetes.io/projected/b9f30537-9153-4862-97aa-31e4f5e57ecf-kube-api-access-g7wwn\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.968809 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.968887 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9f30537-9153-4862-97aa-31e4f5e57ecf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.968976 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.969147 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9f30537-9153-4862-97aa-31e4f5e57ecf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:54 crc kubenswrapper[4669]: I1010 09:26:54.969318 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.071539 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9f30537-9153-4862-97aa-31e4f5e57ecf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.071643 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.072360 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.072427 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f30537-9153-4862-97aa-31e4f5e57ecf-config\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.072462 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7wwn\" (UniqueName: \"kubernetes.io/projected/b9f30537-9153-4862-97aa-31e4f5e57ecf-kube-api-access-g7wwn\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.072534 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.072557 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9f30537-9153-4862-97aa-31e4f5e57ecf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.072607 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.072872 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.073321 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9f30537-9153-4862-97aa-31e4f5e57ecf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.073813 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9f30537-9153-4862-97aa-31e4f5e57ecf-config\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.074190 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9f30537-9153-4862-97aa-31e4f5e57ecf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.077598 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.081172 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.090063 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7wwn\" (UniqueName: \"kubernetes.io/projected/b9f30537-9153-4862-97aa-31e4f5e57ecf-kube-api-access-g7wwn\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.091856 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9f30537-9153-4862-97aa-31e4f5e57ecf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.093146 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b9f30537-9153-4862-97aa-31e4f5e57ecf\") " pod="openstack/ovsdbserver-sb-0" Oct 10 09:26:55 crc kubenswrapper[4669]: I1010 09:26:55.244201 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.959830 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.960708 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rxqbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-qsllq_openstack(1cccc3a1-eac2-4afe-b25c-27d03564ba8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.962001 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-qsllq" podUID="1cccc3a1-eac2-4afe-b25c-27d03564ba8d" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.962283 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.962666 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-78vkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-77597f887-6jmnh_openstack(eafc1e37-eb15-4615-b155-1eb20365e0b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.963849 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-77597f887-6jmnh" podUID="eafc1e37-eb15-4615-b155-1eb20365e0b5" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.987967 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.988144 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j7d8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8575fc99d7-f75sk_openstack(739d7b10-7e0a-4d33-9be2-2c6676088dbc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:27:08 crc kubenswrapper[4669]: E1010 09:27:08.990966 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" podUID="739d7b10-7e0a-4d33-9be2-2c6676088dbc" Oct 10 09:27:09 crc kubenswrapper[4669]: E1010 09:27:09.052415 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 10 09:27:09 crc kubenswrapper[4669]: E1010 09:27:09.052548 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-57fz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-6xz9c_openstack(b3f05e5f-8bd1-4a18-b789-be400ccfa7e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:27:09 crc kubenswrapper[4669]: E1010 09:27:09.053747 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" podUID="b3f05e5f-8bd1-4a18-b789-be400ccfa7e5" Oct 10 09:27:09 crc kubenswrapper[4669]: I1010 09:27:09.258287 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jrw5q"] Oct 10 09:27:09 crc kubenswrapper[4669]: W1010 09:27:09.448501 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b815c31_96c7_47c3_b458_35945b55b326.slice/crio-602b685eebc51555dd2a0216e78b433ac93f9929c4405d119cb06cb78fe488a8 WatchSource:0}: Error finding container 602b685eebc51555dd2a0216e78b433ac93f9929c4405d119cb06cb78fe488a8: Status 404 returned error can't find the container with id 602b685eebc51555dd2a0216e78b433ac93f9929c4405d119cb06cb78fe488a8 Oct 10 09:27:09 crc kubenswrapper[4669]: I1010 09:27:09.698237 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 09:27:09 crc kubenswrapper[4669]: I1010 09:27:09.991644 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q" event={"ID":"2b815c31-96c7-47c3-b458-35945b55b326","Type":"ContainerStarted","Data":"602b685eebc51555dd2a0216e78b433ac93f9929c4405d119cb06cb78fe488a8"} Oct 10 09:27:09 crc kubenswrapper[4669]: I1010 09:27:09.994163 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef"} Oct 10 09:27:09 crc kubenswrapper[4669]: I1010 09:27:09.995898 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9c376674-4cf1-40c0-9e3c-c25e3fefad30","Type":"ContainerStarted","Data":"9d06237d56b9ca22a9c9f54b17109377fe5c74472ae428cba3ebb130846322dc"} Oct 10 09:27:09 crc kubenswrapper[4669]: E1010 09:27:09.997336 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" podUID="739d7b10-7e0a-4d33-9be2-2c6676088dbc" Oct 10 09:27:09 crc kubenswrapper[4669]: E1010 09:27:09.998731 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-77597f887-6jmnh" podUID="eafc1e37-eb15-4615-b155-1eb20365e0b5" Oct 10 09:27:10 crc kubenswrapper[4669]: E1010 09:27:10.063993 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 10 09:27:10 crc kubenswrapper[4669]: E1010 09:27:10.064044 4669 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 10 09:27:10 crc kubenswrapper[4669]: E1010 09:27:10.064185 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2l9k8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:27:10 crc kubenswrapper[4669]: E1010 09:27:10.072500 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.073684 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=4.371160941 podStartE2EDuration="25.073672305s" podCreationTimestamp="2025-10-10 09:26:45 +0000 UTC" firstStartedPulling="2025-10-10 09:26:47.340718144 +0000 UTC m=+950.356736886" lastFinishedPulling="2025-10-10 09:27:08.043229468 +0000 UTC m=+971.059248250" observedRunningTime="2025-10-10 09:27:10.0409849 +0000 UTC m=+973.057003652" watchObservedRunningTime="2025-10-10 09:27:10.073672305 +0000 UTC m=+973.089691047" Oct 10 09:27:10 crc kubenswrapper[4669]: W1010 09:27:10.090206 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9f30537_9153_4862_97aa_31e4f5e57ecf.slice/crio-9c3d79ad75680f309fdf38c9d2bd6db50061d84fda5eef338901d55ea6d0c19b WatchSource:0}: Error finding container 9c3d79ad75680f309fdf38c9d2bd6db50061d84fda5eef338901d55ea6d0c19b: Status 404 returned error can't find the container with id 9c3d79ad75680f309fdf38c9d2bd6db50061d84fda5eef338901d55ea6d0c19b Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.319867 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.428962 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57fz7\" (UniqueName: \"kubernetes.io/projected/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-kube-api-access-57fz7\") pod \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.429716 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-config\") pod \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\" (UID: \"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5\") " Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.430521 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-config" (OuterVolumeSpecName: "config") pod "b3f05e5f-8bd1-4a18-b789-be400ccfa7e5" (UID: "b3f05e5f-8bd1-4a18-b789-be400ccfa7e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.431489 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.462826 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-kube-api-access-57fz7" (OuterVolumeSpecName: "kube-api-access-57fz7") pod "b3f05e5f-8bd1-4a18-b789-be400ccfa7e5" (UID: "b3f05e5f-8bd1-4a18-b789-be400ccfa7e5"). InnerVolumeSpecName "kube-api-access-57fz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.468527 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.533057 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxqbv\" (UniqueName: \"kubernetes.io/projected/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-kube-api-access-rxqbv\") pod \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.533129 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-dns-svc\") pod \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.533327 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-config\") pod \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\" (UID: \"1cccc3a1-eac2-4afe-b25c-27d03564ba8d\") " Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.533569 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1cccc3a1-eac2-4afe-b25c-27d03564ba8d" (UID: "1cccc3a1-eac2-4afe-b25c-27d03564ba8d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.533955 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.533980 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57fz7\" (UniqueName: \"kubernetes.io/projected/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5-kube-api-access-57fz7\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.534023 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-config" (OuterVolumeSpecName: "config") pod "1cccc3a1-eac2-4afe-b25c-27d03564ba8d" (UID: "1cccc3a1-eac2-4afe-b25c-27d03564ba8d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.535240 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-kube-api-access-rxqbv" (OuterVolumeSpecName: "kube-api-access-rxqbv") pod "1cccc3a1-eac2-4afe-b25c-27d03564ba8d" (UID: "1cccc3a1-eac2-4afe-b25c-27d03564ba8d"). InnerVolumeSpecName "kube-api-access-rxqbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.635727 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.635762 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxqbv\" (UniqueName: \"kubernetes.io/projected/1cccc3a1-eac2-4afe-b25c-27d03564ba8d-kube-api-access-rxqbv\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.689677 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-d8wg9"] Oct 10 09:27:10 crc kubenswrapper[4669]: I1010 09:27:10.789347 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.006216 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-d8wg9" event={"ID":"db075709-6fea-46c4-b711-c5517eaae350","Type":"ContainerStarted","Data":"0a4bc9f3b6edfb39a722d20eda7e728b1ca449e2f5c0a9453cc5625a2ba3aa07"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.007274 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b9f30537-9153-4862-97aa-31e4f5e57ecf","Type":"ContainerStarted","Data":"9c3d79ad75680f309fdf38c9d2bd6db50061d84fda5eef338901d55ea6d0c19b"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.008456 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"79d402e9-8187-4cd5-9df5-9297d8fc4b5a","Type":"ContainerStarted","Data":"0e074fdebd3330234a6bfa3737c2430986dbd38d05dcbdac3be0e8a123506f1e"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.010657 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5321fb79-5d79-4094-b314-8914ed41cfd5","Type":"ContainerStarted","Data":"8fd43e8d63f0706cc052de83303a18e29c3061d753dc6ff0439b37f9dbac7600"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.021078 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"02229a1f-8349-4e0c-b5b5-b56ba246f395","Type":"ContainerStarted","Data":"532312a1a439199737a5d8daf71c58276bc338bdc74ccea11fc900b09c69d76c"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.022954 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-qsllq" event={"ID":"1cccc3a1-eac2-4afe-b25c-27d03564ba8d","Type":"ContainerDied","Data":"493937a410540711d6a7bc825978e109521008bb2a86ec8080e31906b47b5de1"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.022985 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-qsllq" Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.023983 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" event={"ID":"b3f05e5f-8bd1-4a18-b789-be400ccfa7e5","Type":"ContainerDied","Data":"b369b9773bfa36be506c34ee10a84ac50555ebb296212f7dd34db2994d26284f"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.024021 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-6xz9c" Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.029508 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b787ed5f-cbcf-4144-a87f-a014dbfd5282","Type":"ContainerStarted","Data":"971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3"} Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.032358 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 10 09:27:11 crc kubenswrapper[4669]: E1010 09:27:11.040696 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.168303 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-6xz9c"] Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.172719 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-6xz9c"] Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.208595 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qsllq"] Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.215337 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-qsllq"] Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.805500 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cccc3a1-eac2-4afe-b25c-27d03564ba8d" path="/var/lib/kubelet/pods/1cccc3a1-eac2-4afe-b25c-27d03564ba8d/volumes" Oct 10 09:27:11 crc kubenswrapper[4669]: I1010 09:27:11.805877 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3f05e5f-8bd1-4a18-b789-be400ccfa7e5" path="/var/lib/kubelet/pods/b3f05e5f-8bd1-4a18-b789-be400ccfa7e5/volumes" Oct 10 09:27:12 crc kubenswrapper[4669]: I1010 09:27:12.039821 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c46f84af-548a-4390-8242-c48ff4fc25aa","Type":"ContainerStarted","Data":"c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8"} Oct 10 09:27:15 crc kubenswrapper[4669]: I1010 09:27:15.072759 4669 generic.go:334] "Generic (PLEG): container finished" podID="79d402e9-8187-4cd5-9df5-9297d8fc4b5a" containerID="0e074fdebd3330234a6bfa3737c2430986dbd38d05dcbdac3be0e8a123506f1e" exitCode=0 Oct 10 09:27:15 crc kubenswrapper[4669]: I1010 09:27:15.073057 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"79d402e9-8187-4cd5-9df5-9297d8fc4b5a","Type":"ContainerDied","Data":"0e074fdebd3330234a6bfa3737c2430986dbd38d05dcbdac3be0e8a123506f1e"} Oct 10 09:27:15 crc kubenswrapper[4669]: I1010 09:27:15.074956 4669 generic.go:334] "Generic (PLEG): container finished" podID="02229a1f-8349-4e0c-b5b5-b56ba246f395" containerID="532312a1a439199737a5d8daf71c58276bc338bdc74ccea11fc900b09c69d76c" exitCode=0 Oct 10 09:27:15 crc kubenswrapper[4669]: I1010 09:27:15.074998 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"02229a1f-8349-4e0c-b5b5-b56ba246f395","Type":"ContainerDied","Data":"532312a1a439199737a5d8daf71c58276bc338bdc74ccea11fc900b09c69d76c"} Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.086260 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q" event={"ID":"2b815c31-96c7-47c3-b458-35945b55b326","Type":"ContainerStarted","Data":"fbedefb8ab82d6de513e293bd1bcfef0c821657df7a4044f9da32dc6ce4ff30b"} Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.087484 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jrw5q" Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.090455 4669 generic.go:334] "Generic (PLEG): container finished" podID="db075709-6fea-46c4-b711-c5517eaae350" containerID="e92c92a01f02012206a9a82e02e1b2863f235606fe46b3ae353b285f764d67e8" exitCode=0 Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.090555 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-d8wg9" event={"ID":"db075709-6fea-46c4-b711-c5517eaae350","Type":"ContainerDied","Data":"e92c92a01f02012206a9a82e02e1b2863f235606fe46b3ae353b285f764d67e8"} Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.098039 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b9f30537-9153-4862-97aa-31e4f5e57ecf","Type":"ContainerStarted","Data":"c718ed5dfea2a1dc82c93f0ded34e5fedd88fcde7eeff676e29cec9ce6931076"} Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.103064 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"79d402e9-8187-4cd5-9df5-9297d8fc4b5a","Type":"ContainerStarted","Data":"c99e2ea0445ddb124885e4c0bd8d0dab48384671091bd460fdc664d41a31055d"} Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.109930 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5321fb79-5d79-4094-b314-8914ed41cfd5","Type":"ContainerStarted","Data":"efccc86b6dc3f45b836dc39a4a8af0d6c0ef62214cf5305f6f9c213aee18f0f6"} Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.112957 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"02229a1f-8349-4e0c-b5b5-b56ba246f395","Type":"ContainerStarted","Data":"34c9e0135114c2debe676b771773e9419e74bdbcef2be58064747fc6b3ebf30c"} Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.115499 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jrw5q" podStartSLOduration=20.882362558 podStartE2EDuration="26.115486324s" podCreationTimestamp="2025-10-10 09:26:50 +0000 UTC" firstStartedPulling="2025-10-10 09:27:09.451171689 +0000 UTC m=+972.467190431" lastFinishedPulling="2025-10-10 09:27:14.684295445 +0000 UTC m=+977.700314197" observedRunningTime="2025-10-10 09:27:16.107532467 +0000 UTC m=+979.123551209" watchObservedRunningTime="2025-10-10 09:27:16.115486324 +0000 UTC m=+979.131505066" Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.129927 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.112061334 podStartE2EDuration="33.129912949s" podCreationTimestamp="2025-10-10 09:26:43 +0000 UTC" firstStartedPulling="2025-10-10 09:26:45.209907186 +0000 UTC m=+948.225925928" lastFinishedPulling="2025-10-10 09:27:09.2277588 +0000 UTC m=+972.243777543" observedRunningTime="2025-10-10 09:27:16.12929249 +0000 UTC m=+979.145311222" watchObservedRunningTime="2025-10-10 09:27:16.129912949 +0000 UTC m=+979.145931691" Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.170373 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.580167772 podStartE2EDuration="32.170350204s" podCreationTimestamp="2025-10-10 09:26:44 +0000 UTC" firstStartedPulling="2025-10-10 09:26:47.185076384 +0000 UTC m=+950.201095136" lastFinishedPulling="2025-10-10 09:27:08.775258816 +0000 UTC m=+971.791277568" observedRunningTime="2025-10-10 09:27:16.167303166 +0000 UTC m=+979.183321918" watchObservedRunningTime="2025-10-10 09:27:16.170350204 +0000 UTC m=+979.186368956" Oct 10 09:27:16 crc kubenswrapper[4669]: I1010 09:27:16.224797 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 10 09:27:17 crc kubenswrapper[4669]: I1010 09:27:17.126848 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-d8wg9" event={"ID":"db075709-6fea-46c4-b711-c5517eaae350","Type":"ContainerStarted","Data":"8ea942f3de23c5c7899ceb293d5e3381d88f1d396be8dfd389520244bbf3c6c9"} Oct 10 09:27:17 crc kubenswrapper[4669]: I1010 09:27:17.127166 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-d8wg9" event={"ID":"db075709-6fea-46c4-b711-c5517eaae350","Type":"ContainerStarted","Data":"26414afda5d63d0b59000729f539456b553f21d812a9f6c8b28d679ed49898e9"} Oct 10 09:27:17 crc kubenswrapper[4669]: I1010 09:27:17.127529 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:27:17 crc kubenswrapper[4669]: I1010 09:27:17.826767 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-d8wg9" podStartSLOduration=23.872705156 podStartE2EDuration="27.82675023s" podCreationTimestamp="2025-10-10 09:26:50 +0000 UTC" firstStartedPulling="2025-10-10 09:27:10.693671 +0000 UTC m=+973.709689742" lastFinishedPulling="2025-10-10 09:27:14.647716074 +0000 UTC m=+977.663734816" observedRunningTime="2025-10-10 09:27:17.14800907 +0000 UTC m=+980.164027802" watchObservedRunningTime="2025-10-10 09:27:17.82675023 +0000 UTC m=+980.842768972" Oct 10 09:27:18 crc kubenswrapper[4669]: I1010 09:27:18.133834 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.142408 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b9f30537-9153-4862-97aa-31e4f5e57ecf","Type":"ContainerStarted","Data":"a2dc11f2e2c2f97e4f5d4d3dbb5a5aa7b4acd9d6f35f747abcc4268de0fbcaa6"} Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.144853 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5321fb79-5d79-4094-b314-8914ed41cfd5","Type":"ContainerStarted","Data":"35b67af466e28899afe4117f9b8f0c75fb118a0b4d3c21e48c134c3ed9758a3a"} Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.168481 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=17.704143837 podStartE2EDuration="26.168460123s" podCreationTimestamp="2025-10-10 09:26:53 +0000 UTC" firstStartedPulling="2025-10-10 09:27:10.103561299 +0000 UTC m=+973.119580041" lastFinishedPulling="2025-10-10 09:27:18.567877585 +0000 UTC m=+981.583896327" observedRunningTime="2025-10-10 09:27:19.162943255 +0000 UTC m=+982.178961997" watchObservedRunningTime="2025-10-10 09:27:19.168460123 +0000 UTC m=+982.184478875" Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.244949 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.290824 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.312863 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=20.617757296 podStartE2EDuration="28.312841341s" podCreationTimestamp="2025-10-10 09:26:51 +0000 UTC" firstStartedPulling="2025-10-10 09:27:10.878204144 +0000 UTC m=+973.894222886" lastFinishedPulling="2025-10-10 09:27:18.573288189 +0000 UTC m=+981.589306931" observedRunningTime="2025-10-10 09:27:19.189513793 +0000 UTC m=+982.205532555" watchObservedRunningTime="2025-10-10 09:27:19.312841341 +0000 UTC m=+982.328860083" Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.788291 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 10 09:27:19 crc kubenswrapper[4669]: I1010 09:27:19.825354 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.152146 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.152213 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.219957 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.223001 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.412993 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-f75sk"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.456854 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vg62q"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.464026 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.477889 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.512062 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-24xjg"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.515595 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.517214 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.526738 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2fdb85-cfc4-471d-b35e-9388c3406895-config\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.526803 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjh2k\" (UniqueName: \"kubernetes.io/projected/1f2fdb85-cfc4-471d-b35e-9388c3406895-kube-api-access-sjh2k\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.526840 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f2fdb85-cfc4-471d-b35e-9388c3406895-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.526880 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1f2fdb85-cfc4-471d-b35e-9388c3406895-ovs-rundir\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.527030 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2fdb85-cfc4-471d-b35e-9388c3406895-combined-ca-bundle\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.527260 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1f2fdb85-cfc4-471d-b35e-9388c3406895-ovn-rundir\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.567574 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-24xjg"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.594936 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vg62q"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631418 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjh2k\" (UniqueName: \"kubernetes.io/projected/1f2fdb85-cfc4-471d-b35e-9388c3406895-kube-api-access-sjh2k\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-ovsdbserver-sb\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631495 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-config\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631513 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f2fdb85-cfc4-471d-b35e-9388c3406895-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631532 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-dns-svc\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631560 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1f2fdb85-cfc4-471d-b35e-9388c3406895-ovs-rundir\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631599 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsz2g\" (UniqueName: \"kubernetes.io/projected/394ab4cc-4c59-4881-b592-4782895fd278-kube-api-access-fsz2g\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631628 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2fdb85-cfc4-471d-b35e-9388c3406895-combined-ca-bundle\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631662 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1f2fdb85-cfc4-471d-b35e-9388c3406895-ovn-rundir\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.631693 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2fdb85-cfc4-471d-b35e-9388c3406895-config\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.632329 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2fdb85-cfc4-471d-b35e-9388c3406895-config\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.634269 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-6jmnh"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.635028 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1f2fdb85-cfc4-471d-b35e-9388c3406895-ovs-rundir\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.635084 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1f2fdb85-cfc4-471d-b35e-9388c3406895-ovn-rundir\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.645629 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2fdb85-cfc4-471d-b35e-9388c3406895-combined-ca-bundle\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.651903 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f2fdb85-cfc4-471d-b35e-9388c3406895-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.701624 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjh2k\" (UniqueName: \"kubernetes.io/projected/1f2fdb85-cfc4-471d-b35e-9388c3406895-kube-api-access-sjh2k\") pod \"ovn-controller-metrics-vg62q\" (UID: \"1f2fdb85-cfc4-471d-b35e-9388c3406895\") " pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.733848 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsz2g\" (UniqueName: \"kubernetes.io/projected/394ab4cc-4c59-4881-b592-4782895fd278-kube-api-access-fsz2g\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.734016 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-ovsdbserver-sb\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.734039 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-config\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.734064 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-dns-svc\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.735428 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-ovsdbserver-sb\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.736256 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-dns-svc\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.736276 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-config\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.756858 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsz2g\" (UniqueName: \"kubernetes.io/projected/394ab4cc-4c59-4881-b592-4782895fd278-kube-api-access-fsz2g\") pod \"dnsmasq-dns-545fb8c44f-24xjg\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.756911 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-9bvgm"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.758184 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.772045 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.836453 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.836505 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.836526 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khpkp\" (UniqueName: \"kubernetes.io/projected/a5bce784-8452-4b77-a1d0-539e471aed1c-kube-api-access-khpkp\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.836551 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.836571 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-config\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.838069 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-9bvgm"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.846568 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.847059 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vg62q" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.847843 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.858278 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.861945 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.862267 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.862812 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-nm8s9" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.870629 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.870794 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937539 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937597 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937619 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937636 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khpkp\" (UniqueName: \"kubernetes.io/projected/a5bce784-8452-4b77-a1d0-539e471aed1c-kube-api-access-khpkp\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937651 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/028c96a4-38cf-4710-bcb2-a132836d77f1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937675 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937702 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/028c96a4-38cf-4710-bcb2-a132836d77f1-scripts\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937717 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937732 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-config\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937793 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2sdj\" (UniqueName: \"kubernetes.io/projected/028c96a4-38cf-4710-bcb2-a132836d77f1-kube-api-access-s2sdj\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937813 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028c96a4-38cf-4710-bcb2-a132836d77f1-config\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.937848 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.938654 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.939146 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.940647 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-config\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.950143 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:20 crc kubenswrapper[4669]: I1010 09:27:20.979035 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.014614 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khpkp\" (UniqueName: \"kubernetes.io/projected/a5bce784-8452-4b77-a1d0-539e471aed1c-kube-api-access-khpkp\") pod \"dnsmasq-dns-dc9d58d7-9bvgm\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.046183 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7d8k\" (UniqueName: \"kubernetes.io/projected/739d7b10-7e0a-4d33-9be2-2c6676088dbc-kube-api-access-j7d8k\") pod \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.046267 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-dns-svc\") pod \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.046951 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-config\") pod \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\" (UID: \"739d7b10-7e0a-4d33-9be2-2c6676088dbc\") " Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047320 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047352 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047347 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "739d7b10-7e0a-4d33-9be2-2c6676088dbc" (UID: "739d7b10-7e0a-4d33-9be2-2c6676088dbc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047381 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/028c96a4-38cf-4710-bcb2-a132836d77f1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047412 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/028c96a4-38cf-4710-bcb2-a132836d77f1-scripts\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047431 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047494 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2sdj\" (UniqueName: \"kubernetes.io/projected/028c96a4-38cf-4710-bcb2-a132836d77f1-kube-api-access-s2sdj\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047515 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028c96a4-38cf-4710-bcb2-a132836d77f1-config\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.047570 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.048222 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/028c96a4-38cf-4710-bcb2-a132836d77f1-config\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.051165 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/028c96a4-38cf-4710-bcb2-a132836d77f1-scripts\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.051406 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/028c96a4-38cf-4710-bcb2-a132836d77f1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.055753 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.058001 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/739d7b10-7e0a-4d33-9be2-2c6676088dbc-kube-api-access-j7d8k" (OuterVolumeSpecName: "kube-api-access-j7d8k") pod "739d7b10-7e0a-4d33-9be2-2c6676088dbc" (UID: "739d7b10-7e0a-4d33-9be2-2c6676088dbc"). InnerVolumeSpecName "kube-api-access-j7d8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.058618 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.058835 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-config" (OuterVolumeSpecName: "config") pod "739d7b10-7e0a-4d33-9be2-2c6676088dbc" (UID: "739d7b10-7e0a-4d33-9be2-2c6676088dbc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.070096 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/028c96a4-38cf-4710-bcb2-a132836d77f1-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.076506 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2sdj\" (UniqueName: \"kubernetes.io/projected/028c96a4-38cf-4710-bcb2-a132836d77f1-kube-api-access-s2sdj\") pod \"ovn-northd-0\" (UID: \"028c96a4-38cf-4710-bcb2-a132836d77f1\") " pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.106944 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.154721 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7d8k\" (UniqueName: \"kubernetes.io/projected/739d7b10-7e0a-4d33-9be2-2c6676088dbc-kube-api-access-j7d8k\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.154745 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739d7b10-7e0a-4d33-9be2-2c6676088dbc-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.172824 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.173195 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-6jmnh" event={"ID":"eafc1e37-eb15-4615-b155-1eb20365e0b5","Type":"ContainerDied","Data":"c7449585cfbf9681d7d362b3d4b5dc6ceef2f5acd32d6a888e6c319d7ee8b52f"} Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.177071 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.178864 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.181672 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8575fc99d7-f75sk" event={"ID":"739d7b10-7e0a-4d33-9be2-2c6676088dbc","Type":"ContainerDied","Data":"9226a2a596aa547759e7c97637d0c59259eab9bdf03653a51381e055e73cd494"} Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.257053 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-config\") pod \"eafc1e37-eb15-4615-b155-1eb20365e0b5\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.257874 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-dns-svc\") pod \"eafc1e37-eb15-4615-b155-1eb20365e0b5\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.259357 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-config" (OuterVolumeSpecName: "config") pod "eafc1e37-eb15-4615-b155-1eb20365e0b5" (UID: "eafc1e37-eb15-4615-b155-1eb20365e0b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.259780 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eafc1e37-eb15-4615-b155-1eb20365e0b5" (UID: "eafc1e37-eb15-4615-b155-1eb20365e0b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.260537 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-f75sk"] Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.266034 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8575fc99d7-f75sk"] Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.360315 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78vkp\" (UniqueName: \"kubernetes.io/projected/eafc1e37-eb15-4615-b155-1eb20365e0b5-kube-api-access-78vkp\") pod \"eafc1e37-eb15-4615-b155-1eb20365e0b5\" (UID: \"eafc1e37-eb15-4615-b155-1eb20365e0b5\") " Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.360980 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.361000 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eafc1e37-eb15-4615-b155-1eb20365e0b5-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.364806 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eafc1e37-eb15-4615-b155-1eb20365e0b5-kube-api-access-78vkp" (OuterVolumeSpecName: "kube-api-access-78vkp") pod "eafc1e37-eb15-4615-b155-1eb20365e0b5" (UID: "eafc1e37-eb15-4615-b155-1eb20365e0b5"). InnerVolumeSpecName "kube-api-access-78vkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.464188 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78vkp\" (UniqueName: \"kubernetes.io/projected/eafc1e37-eb15-4615-b155-1eb20365e0b5-kube-api-access-78vkp\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:21 crc kubenswrapper[4669]: W1010 09:27:21.630102 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f2fdb85_cfc4_471d_b35e_9388c3406895.slice/crio-53537ce6c3aca1c1b888ca32864d1d867758d3c8bc6dc74c8358b8eb05b4222d WatchSource:0}: Error finding container 53537ce6c3aca1c1b888ca32864d1d867758d3c8bc6dc74c8358b8eb05b4222d: Status 404 returned error can't find the container with id 53537ce6c3aca1c1b888ca32864d1d867758d3c8bc6dc74c8358b8eb05b4222d Oct 10 09:27:21 crc kubenswrapper[4669]: W1010 09:27:21.631972 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod394ab4cc_4c59_4881_b592_4782895fd278.slice/crio-636d89551865b67d5e7b42fbcea5de972b7a0285b75d0b5fdb7fd63f484389b5 WatchSource:0}: Error finding container 636d89551865b67d5e7b42fbcea5de972b7a0285b75d0b5fdb7fd63f484389b5: Status 404 returned error can't find the container with id 636d89551865b67d5e7b42fbcea5de972b7a0285b75d0b5fdb7fd63f484389b5 Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.635001 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-24xjg"] Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.641840 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vg62q"] Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.769139 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-9bvgm"] Oct 10 09:27:21 crc kubenswrapper[4669]: W1010 09:27:21.784843 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5bce784_8452_4b77_a1d0_539e471aed1c.slice/crio-981ba3aebe4610cf6d68bc6bd0fd07d06350a0ad7d989579e2a95a460e807765 WatchSource:0}: Error finding container 981ba3aebe4610cf6d68bc6bd0fd07d06350a0ad7d989579e2a95a460e807765: Status 404 returned error can't find the container with id 981ba3aebe4610cf6d68bc6bd0fd07d06350a0ad7d989579e2a95a460e807765 Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.809231 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="739d7b10-7e0a-4d33-9be2-2c6676088dbc" path="/var/lib/kubelet/pods/739d7b10-7e0a-4d33-9be2-2c6676088dbc/volumes" Oct 10 09:27:21 crc kubenswrapper[4669]: I1010 09:27:21.839477 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.186137 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"028c96a4-38cf-4710-bcb2-a132836d77f1","Type":"ContainerStarted","Data":"2a177588c9d34d9af32f9d197d78dc36a3812d56dc2f6e2b1140d78613c8d93a"} Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.187202 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" event={"ID":"a5bce784-8452-4b77-a1d0-539e471aed1c","Type":"ContainerStarted","Data":"981ba3aebe4610cf6d68bc6bd0fd07d06350a0ad7d989579e2a95a460e807765"} Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.189011 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vg62q" event={"ID":"1f2fdb85-cfc4-471d-b35e-9388c3406895","Type":"ContainerStarted","Data":"59373e999b0e8a59c4def45773d5fb509dbdd8d87d6b32ebee4eaa59c6c4030c"} Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.189035 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vg62q" event={"ID":"1f2fdb85-cfc4-471d-b35e-9388c3406895","Type":"ContainerStarted","Data":"53537ce6c3aca1c1b888ca32864d1d867758d3c8bc6dc74c8358b8eb05b4222d"} Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.191176 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" event={"ID":"394ab4cc-4c59-4881-b592-4782895fd278","Type":"ContainerStarted","Data":"636d89551865b67d5e7b42fbcea5de972b7a0285b75d0b5fdb7fd63f484389b5"} Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.191335 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-6jmnh" Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.213082 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vg62q" podStartSLOduration=2.213061862 podStartE2EDuration="2.213061862s" podCreationTimestamp="2025-10-10 09:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:27:22.208113532 +0000 UTC m=+985.224132284" watchObservedRunningTime="2025-10-10 09:27:22.213061862 +0000 UTC m=+985.229080604" Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.288899 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-6jmnh"] Oct 10 09:27:22 crc kubenswrapper[4669]: I1010 09:27:22.304115 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-6jmnh"] Oct 10 09:27:23 crc kubenswrapper[4669]: I1010 09:27:23.199428 4669 generic.go:334] "Generic (PLEG): container finished" podID="394ab4cc-4c59-4881-b592-4782895fd278" containerID="7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7" exitCode=0 Oct 10 09:27:23 crc kubenswrapper[4669]: I1010 09:27:23.199949 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" event={"ID":"394ab4cc-4c59-4881-b592-4782895fd278","Type":"ContainerDied","Data":"7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7"} Oct 10 09:27:23 crc kubenswrapper[4669]: I1010 09:27:23.806140 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eafc1e37-eb15-4615-b155-1eb20365e0b5" path="/var/lib/kubelet/pods/eafc1e37-eb15-4615-b155-1eb20365e0b5/volumes" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.211567 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"028c96a4-38cf-4710-bcb2-a132836d77f1","Type":"ContainerStarted","Data":"1e40f2adff9f33172d99e035dfcd53dd8ccc1e91db5ac14c99b5d4ff3aa15bcd"} Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.211654 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"028c96a4-38cf-4710-bcb2-a132836d77f1","Type":"ContainerStarted","Data":"d6b06feb986bb22676de41dc2014fc3f96d8f62df4c13b064606e1953808c50f"} Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.211723 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.213915 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79","Type":"ContainerStarted","Data":"1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34"} Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.214174 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.215729 4669 generic.go:334] "Generic (PLEG): container finished" podID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerID="5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e" exitCode=0 Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.215788 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" event={"ID":"a5bce784-8452-4b77-a1d0-539e471aed1c","Type":"ContainerDied","Data":"5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e"} Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.218163 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" event={"ID":"394ab4cc-4c59-4881-b592-4782895fd278","Type":"ContainerStarted","Data":"4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e"} Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.218480 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.254083 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.013611262 podStartE2EDuration="4.254060248s" podCreationTimestamp="2025-10-10 09:27:20 +0000 UTC" firstStartedPulling="2025-10-10 09:27:21.8473161 +0000 UTC m=+984.863334842" lastFinishedPulling="2025-10-10 09:27:23.087765086 +0000 UTC m=+986.103783828" observedRunningTime="2025-10-10 09:27:24.247992653 +0000 UTC m=+987.264011395" watchObservedRunningTime="2025-10-10 09:27:24.254060248 +0000 UTC m=+987.270079010" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.280302 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" podStartSLOduration=3.495929185 podStartE2EDuration="4.280285024s" podCreationTimestamp="2025-10-10 09:27:20 +0000 UTC" firstStartedPulling="2025-10-10 09:27:21.633973716 +0000 UTC m=+984.649992458" lastFinishedPulling="2025-10-10 09:27:22.418329545 +0000 UTC m=+985.434348297" observedRunningTime="2025-10-10 09:27:24.279433086 +0000 UTC m=+987.295451838" watchObservedRunningTime="2025-10-10 09:27:24.280285024 +0000 UTC m=+987.296303776" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.331593 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.643779097 podStartE2EDuration="37.331558399s" podCreationTimestamp="2025-10-10 09:26:47 +0000 UTC" firstStartedPulling="2025-10-10 09:26:48.5439633 +0000 UTC m=+951.559982042" lastFinishedPulling="2025-10-10 09:27:23.231742592 +0000 UTC m=+986.247761344" observedRunningTime="2025-10-10 09:27:24.323525 +0000 UTC m=+987.339543742" watchObservedRunningTime="2025-10-10 09:27:24.331558399 +0000 UTC m=+987.347577141" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.517341 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.517857 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 10 09:27:24 crc kubenswrapper[4669]: I1010 09:27:24.583906 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.231502 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" event={"ID":"a5bce784-8452-4b77-a1d0-539e471aed1c","Type":"ContainerStarted","Data":"91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b"} Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.231550 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.258521 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" podStartSLOduration=4.019290614 podStartE2EDuration="5.258501048s" podCreationTimestamp="2025-10-10 09:27:20 +0000 UTC" firstStartedPulling="2025-10-10 09:27:21.786816919 +0000 UTC m=+984.802835661" lastFinishedPulling="2025-10-10 09:27:23.026027353 +0000 UTC m=+986.042046095" observedRunningTime="2025-10-10 09:27:25.253446125 +0000 UTC m=+988.269464877" watchObservedRunningTime="2025-10-10 09:27:25.258501048 +0000 UTC m=+988.274519790" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.296527 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.830028 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-l8tnz"] Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.831391 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l8tnz" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.840221 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-l8tnz"] Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.843227 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.843385 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.893213 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 10 09:27:25 crc kubenswrapper[4669]: I1010 09:27:25.945824 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh29z\" (UniqueName: \"kubernetes.io/projected/2ae32e28-6814-4693-b3e0-bdf5d97f9744-kube-api-access-sh29z\") pod \"keystone-db-create-l8tnz\" (UID: \"2ae32e28-6814-4693-b3e0-bdf5d97f9744\") " pod="openstack/keystone-db-create-l8tnz" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.047239 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh29z\" (UniqueName: \"kubernetes.io/projected/2ae32e28-6814-4693-b3e0-bdf5d97f9744-kube-api-access-sh29z\") pod \"keystone-db-create-l8tnz\" (UID: \"2ae32e28-6814-4693-b3e0-bdf5d97f9744\") " pod="openstack/keystone-db-create-l8tnz" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.064876 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh29z\" (UniqueName: \"kubernetes.io/projected/2ae32e28-6814-4693-b3e0-bdf5d97f9744-kube-api-access-sh29z\") pod \"keystone-db-create-l8tnz\" (UID: \"2ae32e28-6814-4693-b3e0-bdf5d97f9744\") " pod="openstack/keystone-db-create-l8tnz" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.153615 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l8tnz" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.231569 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-x8kx4"] Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.233018 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x8kx4" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.238495 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x8kx4"] Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.297761 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.354460 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mncd\" (UniqueName: \"kubernetes.io/projected/30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c-kube-api-access-6mncd\") pod \"placement-db-create-x8kx4\" (UID: \"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c\") " pod="openstack/placement-db-create-x8kx4" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.423783 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8gt4t"] Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.424787 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8gt4t" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.431406 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8gt4t"] Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.461724 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mncd\" (UniqueName: \"kubernetes.io/projected/30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c-kube-api-access-6mncd\") pod \"placement-db-create-x8kx4\" (UID: \"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c\") " pod="openstack/placement-db-create-x8kx4" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.503563 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mncd\" (UniqueName: \"kubernetes.io/projected/30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c-kube-api-access-6mncd\") pod \"placement-db-create-x8kx4\" (UID: \"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c\") " pod="openstack/placement-db-create-x8kx4" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.562845 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgmc4\" (UniqueName: \"kubernetes.io/projected/913a206c-349d-4125-a118-1798385a3d1a-kube-api-access-fgmc4\") pod \"glance-db-create-8gt4t\" (UID: \"913a206c-349d-4125-a118-1798385a3d1a\") " pod="openstack/glance-db-create-8gt4t" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.567863 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x8kx4" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.664618 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgmc4\" (UniqueName: \"kubernetes.io/projected/913a206c-349d-4125-a118-1798385a3d1a-kube-api-access-fgmc4\") pod \"glance-db-create-8gt4t\" (UID: \"913a206c-349d-4125-a118-1798385a3d1a\") " pod="openstack/glance-db-create-8gt4t" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.689414 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgmc4\" (UniqueName: \"kubernetes.io/projected/913a206c-349d-4125-a118-1798385a3d1a-kube-api-access-fgmc4\") pod \"glance-db-create-8gt4t\" (UID: \"913a206c-349d-4125-a118-1798385a3d1a\") " pod="openstack/glance-db-create-8gt4t" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.693485 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-l8tnz"] Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.741879 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8gt4t" Oct 10 09:27:26 crc kubenswrapper[4669]: I1010 09:27:26.861962 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x8kx4"] Oct 10 09:27:26 crc kubenswrapper[4669]: W1010 09:27:26.872445 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30aa5ac3_da2c_4bc6_a82d_cfaab0d0fb8c.slice/crio-43eb96d04bbcbbfaf42b8071dc2aaea1a309226dc76d9fba3d12e78b057e37ed WatchSource:0}: Error finding container 43eb96d04bbcbbfaf42b8071dc2aaea1a309226dc76d9fba3d12e78b057e37ed: Status 404 returned error can't find the container with id 43eb96d04bbcbbfaf42b8071dc2aaea1a309226dc76d9fba3d12e78b057e37ed Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.224909 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8gt4t"] Oct 10 09:27:27 crc kubenswrapper[4669]: W1010 09:27:27.238575 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod913a206c_349d_4125_a118_1798385a3d1a.slice/crio-c45f2d5e9de4d6ac2df013ea8b4fbcc1836fd2889bd9200d3053dc18240d0eb1 WatchSource:0}: Error finding container c45f2d5e9de4d6ac2df013ea8b4fbcc1836fd2889bd9200d3053dc18240d0eb1: Status 404 returned error can't find the container with id c45f2d5e9de4d6ac2df013ea8b4fbcc1836fd2889bd9200d3053dc18240d0eb1 Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.271881 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8gt4t" event={"ID":"913a206c-349d-4125-a118-1798385a3d1a","Type":"ContainerStarted","Data":"c45f2d5e9de4d6ac2df013ea8b4fbcc1836fd2889bd9200d3053dc18240d0eb1"} Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.275804 4669 generic.go:334] "Generic (PLEG): container finished" podID="30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c" containerID="f50c3b443c87455a1f988ec1ca512d2019dfb1a2e8e62c61039ca9547799fdb1" exitCode=0 Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.275975 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x8kx4" event={"ID":"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c","Type":"ContainerDied","Data":"f50c3b443c87455a1f988ec1ca512d2019dfb1a2e8e62c61039ca9547799fdb1"} Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.276023 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x8kx4" event={"ID":"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c","Type":"ContainerStarted","Data":"43eb96d04bbcbbfaf42b8071dc2aaea1a309226dc76d9fba3d12e78b057e37ed"} Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.281673 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l8tnz" event={"ID":"2ae32e28-6814-4693-b3e0-bdf5d97f9744","Type":"ContainerDied","Data":"09df8d6ef49cd88e2f1956c94859dbb324cade7f8699dfadd74f2104497986d5"} Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.281651 4669 generic.go:334] "Generic (PLEG): container finished" podID="2ae32e28-6814-4693-b3e0-bdf5d97f9744" containerID="09df8d6ef49cd88e2f1956c94859dbb324cade7f8699dfadd74f2104497986d5" exitCode=0 Oct 10 09:27:27 crc kubenswrapper[4669]: I1010 09:27:27.282141 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l8tnz" event={"ID":"2ae32e28-6814-4693-b3e0-bdf5d97f9744","Type":"ContainerStarted","Data":"1a69c3efbe0c93985f7a2b1859986d99a966c26b29c452047394e21cb0efe981"} Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.298820 4669 generic.go:334] "Generic (PLEG): container finished" podID="913a206c-349d-4125-a118-1798385a3d1a" containerID="87652d4373bd9770f7bb009fd863ec138287023d6b394c7b0608446578e3c6df" exitCode=0 Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.299919 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8gt4t" event={"ID":"913a206c-349d-4125-a118-1798385a3d1a","Type":"ContainerDied","Data":"87652d4373bd9770f7bb009fd863ec138287023d6b394c7b0608446578e3c6df"} Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.720735 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x8kx4" Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.735023 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l8tnz" Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.802760 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh29z\" (UniqueName: \"kubernetes.io/projected/2ae32e28-6814-4693-b3e0-bdf5d97f9744-kube-api-access-sh29z\") pod \"2ae32e28-6814-4693-b3e0-bdf5d97f9744\" (UID: \"2ae32e28-6814-4693-b3e0-bdf5d97f9744\") " Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.802854 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mncd\" (UniqueName: \"kubernetes.io/projected/30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c-kube-api-access-6mncd\") pod \"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c\" (UID: \"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c\") " Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.815638 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c-kube-api-access-6mncd" (OuterVolumeSpecName: "kube-api-access-6mncd") pod "30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c" (UID: "30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c"). InnerVolumeSpecName "kube-api-access-6mncd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.815774 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae32e28-6814-4693-b3e0-bdf5d97f9744-kube-api-access-sh29z" (OuterVolumeSpecName: "kube-api-access-sh29z") pod "2ae32e28-6814-4693-b3e0-bdf5d97f9744" (UID: "2ae32e28-6814-4693-b3e0-bdf5d97f9744"). InnerVolumeSpecName "kube-api-access-sh29z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.904710 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh29z\" (UniqueName: \"kubernetes.io/projected/2ae32e28-6814-4693-b3e0-bdf5d97f9744-kube-api-access-sh29z\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:28 crc kubenswrapper[4669]: I1010 09:27:28.904739 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mncd\" (UniqueName: \"kubernetes.io/projected/30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c-kube-api-access-6mncd\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.310367 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-l8tnz" event={"ID":"2ae32e28-6814-4693-b3e0-bdf5d97f9744","Type":"ContainerDied","Data":"1a69c3efbe0c93985f7a2b1859986d99a966c26b29c452047394e21cb0efe981"} Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.310440 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a69c3efbe0c93985f7a2b1859986d99a966c26b29c452047394e21cb0efe981" Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.310401 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-l8tnz" Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.313103 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x8kx4" event={"ID":"30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c","Type":"ContainerDied","Data":"43eb96d04bbcbbfaf42b8071dc2aaea1a309226dc76d9fba3d12e78b057e37ed"} Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.313228 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43eb96d04bbcbbfaf42b8071dc2aaea1a309226dc76d9fba3d12e78b057e37ed" Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.313386 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x8kx4" Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.533492 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8gt4t" Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.617508 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgmc4\" (UniqueName: \"kubernetes.io/projected/913a206c-349d-4125-a118-1798385a3d1a-kube-api-access-fgmc4\") pod \"913a206c-349d-4125-a118-1798385a3d1a\" (UID: \"913a206c-349d-4125-a118-1798385a3d1a\") " Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.620572 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913a206c-349d-4125-a118-1798385a3d1a-kube-api-access-fgmc4" (OuterVolumeSpecName: "kube-api-access-fgmc4") pod "913a206c-349d-4125-a118-1798385a3d1a" (UID: "913a206c-349d-4125-a118-1798385a3d1a"). InnerVolumeSpecName "kube-api-access-fgmc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:29 crc kubenswrapper[4669]: I1010 09:27:29.719950 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgmc4\" (UniqueName: \"kubernetes.io/projected/913a206c-349d-4125-a118-1798385a3d1a-kube-api-access-fgmc4\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:30 crc kubenswrapper[4669]: I1010 09:27:30.336719 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8gt4t" event={"ID":"913a206c-349d-4125-a118-1798385a3d1a","Type":"ContainerDied","Data":"c45f2d5e9de4d6ac2df013ea8b4fbcc1836fd2889bd9200d3053dc18240d0eb1"} Oct 10 09:27:30 crc kubenswrapper[4669]: I1010 09:27:30.336774 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c45f2d5e9de4d6ac2df013ea8b4fbcc1836fd2889bd9200d3053dc18240d0eb1" Oct 10 09:27:30 crc kubenswrapper[4669]: I1010 09:27:30.336827 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8gt4t" Oct 10 09:27:30 crc kubenswrapper[4669]: I1010 09:27:30.860788 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.108775 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.171486 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-24xjg"] Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.345890 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" podUID="394ab4cc-4c59-4881-b592-4782895fd278" containerName="dnsmasq-dns" containerID="cri-o://4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e" gracePeriod=10 Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.751848 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.862824 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsz2g\" (UniqueName: \"kubernetes.io/projected/394ab4cc-4c59-4881-b592-4782895fd278-kube-api-access-fsz2g\") pod \"394ab4cc-4c59-4881-b592-4782895fd278\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.863195 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-dns-svc\") pod \"394ab4cc-4c59-4881-b592-4782895fd278\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.863339 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-ovsdbserver-sb\") pod \"394ab4cc-4c59-4881-b592-4782895fd278\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.863398 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-config\") pod \"394ab4cc-4c59-4881-b592-4782895fd278\" (UID: \"394ab4cc-4c59-4881-b592-4782895fd278\") " Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.870341 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/394ab4cc-4c59-4881-b592-4782895fd278-kube-api-access-fsz2g" (OuterVolumeSpecName: "kube-api-access-fsz2g") pod "394ab4cc-4c59-4881-b592-4782895fd278" (UID: "394ab4cc-4c59-4881-b592-4782895fd278"). InnerVolumeSpecName "kube-api-access-fsz2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.902922 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-config" (OuterVolumeSpecName: "config") pod "394ab4cc-4c59-4881-b592-4782895fd278" (UID: "394ab4cc-4c59-4881-b592-4782895fd278"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.904856 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "394ab4cc-4c59-4881-b592-4782895fd278" (UID: "394ab4cc-4c59-4881-b592-4782895fd278"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.905527 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "394ab4cc-4c59-4881-b592-4782895fd278" (UID: "394ab4cc-4c59-4881-b592-4782895fd278"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.964415 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsz2g\" (UniqueName: \"kubernetes.io/projected/394ab4cc-4c59-4881-b592-4782895fd278-kube-api-access-fsz2g\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.964449 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.964459 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:31 crc kubenswrapper[4669]: I1010 09:27:31.964468 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/394ab4cc-4c59-4881-b592-4782895fd278-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.355300 4669 generic.go:334] "Generic (PLEG): container finished" podID="394ab4cc-4c59-4881-b592-4782895fd278" containerID="4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e" exitCode=0 Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.355338 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" event={"ID":"394ab4cc-4c59-4881-b592-4782895fd278","Type":"ContainerDied","Data":"4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e"} Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.355362 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" event={"ID":"394ab4cc-4c59-4881-b592-4782895fd278","Type":"ContainerDied","Data":"636d89551865b67d5e7b42fbcea5de972b7a0285b75d0b5fdb7fd63f484389b5"} Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.355353 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545fb8c44f-24xjg" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.355376 4669 scope.go:117] "RemoveContainer" containerID="4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.374329 4669 scope.go:117] "RemoveContainer" containerID="7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.400868 4669 scope.go:117] "RemoveContainer" containerID="4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.400971 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-24xjg"] Oct 10 09:27:32 crc kubenswrapper[4669]: E1010 09:27:32.401574 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e\": container with ID starting with 4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e not found: ID does not exist" containerID="4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.401661 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e"} err="failed to get container status \"4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e\": rpc error: code = NotFound desc = could not find container \"4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e\": container with ID starting with 4a40b94cdc9da1a7761db87189b2335d6fe9ed12d7711cad27b599be3b144e2e not found: ID does not exist" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.401685 4669 scope.go:117] "RemoveContainer" containerID="7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7" Oct 10 09:27:32 crc kubenswrapper[4669]: E1010 09:27:32.402073 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7\": container with ID starting with 7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7 not found: ID does not exist" containerID="7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.402096 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7"} err="failed to get container status \"7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7\": rpc error: code = NotFound desc = could not find container \"7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7\": container with ID starting with 7b27727a1b66697c61c7aa6bb00f915e8cf32959507d03f40fa9cc9f104e48f7 not found: ID does not exist" Oct 10 09:27:32 crc kubenswrapper[4669]: I1010 09:27:32.405798 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-545fb8c44f-24xjg"] Oct 10 09:27:33 crc kubenswrapper[4669]: I1010 09:27:33.815109 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="394ab4cc-4c59-4881-b592-4782895fd278" path="/var/lib/kubelet/pods/394ab4cc-4c59-4881-b592-4782895fd278/volumes" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879135 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2824-account-create-j2pd4"] Oct 10 09:27:35 crc kubenswrapper[4669]: E1010 09:27:35.879564 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879609 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: E1010 09:27:35.879639 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae32e28-6814-4693-b3e0-bdf5d97f9744" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879653 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae32e28-6814-4693-b3e0-bdf5d97f9744" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: E1010 09:27:35.879672 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913a206c-349d-4125-a118-1798385a3d1a" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879682 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="913a206c-349d-4125-a118-1798385a3d1a" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: E1010 09:27:35.879708 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="394ab4cc-4c59-4881-b592-4782895fd278" containerName="dnsmasq-dns" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879719 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="394ab4cc-4c59-4881-b592-4782895fd278" containerName="dnsmasq-dns" Oct 10 09:27:35 crc kubenswrapper[4669]: E1010 09:27:35.879738 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="394ab4cc-4c59-4881-b592-4782895fd278" containerName="init" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879746 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="394ab4cc-4c59-4881-b592-4782895fd278" containerName="init" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879951 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879973 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="913a206c-349d-4125-a118-1798385a3d1a" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.879985 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae32e28-6814-4693-b3e0-bdf5d97f9744" containerName="mariadb-database-create" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.880002 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="394ab4cc-4c59-4881-b592-4782895fd278" containerName="dnsmasq-dns" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.880697 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2824-account-create-j2pd4" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.883766 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.916108 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2824-account-create-j2pd4"] Oct 10 09:27:35 crc kubenswrapper[4669]: I1010 09:27:35.934628 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfxm9\" (UniqueName: \"kubernetes.io/projected/d33626d7-d6dd-45c0-8442-1617295ec697-kube-api-access-tfxm9\") pod \"keystone-2824-account-create-j2pd4\" (UID: \"d33626d7-d6dd-45c0-8442-1617295ec697\") " pod="openstack/keystone-2824-account-create-j2pd4" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.036362 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfxm9\" (UniqueName: \"kubernetes.io/projected/d33626d7-d6dd-45c0-8442-1617295ec697-kube-api-access-tfxm9\") pod \"keystone-2824-account-create-j2pd4\" (UID: \"d33626d7-d6dd-45c0-8442-1617295ec697\") " pod="openstack/keystone-2824-account-create-j2pd4" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.059569 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfxm9\" (UniqueName: \"kubernetes.io/projected/d33626d7-d6dd-45c0-8442-1617295ec697-kube-api-access-tfxm9\") pod \"keystone-2824-account-create-j2pd4\" (UID: \"d33626d7-d6dd-45c0-8442-1617295ec697\") " pod="openstack/keystone-2824-account-create-j2pd4" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.234526 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2824-account-create-j2pd4" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.241865 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.436224 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ca43-account-create-ts8dc"] Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.438731 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca43-account-create-ts8dc" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.442492 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ca43-account-create-ts8dc"] Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.444057 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.545036 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldc5s\" (UniqueName: \"kubernetes.io/projected/aeca1748-8fbf-4502-acb8-f1a2821a5c2b-kube-api-access-ldc5s\") pod \"placement-ca43-account-create-ts8dc\" (UID: \"aeca1748-8fbf-4502-acb8-f1a2821a5c2b\") " pod="openstack/placement-ca43-account-create-ts8dc" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.614573 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-30cc-account-create-5h92x"] Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.615572 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-30cc-account-create-5h92x" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.617486 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.630955 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-30cc-account-create-5h92x"] Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.646449 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jlkt\" (UniqueName: \"kubernetes.io/projected/00a1f32d-7cb9-403c-91aa-28f749dbbc9a-kube-api-access-2jlkt\") pod \"glance-30cc-account-create-5h92x\" (UID: \"00a1f32d-7cb9-403c-91aa-28f749dbbc9a\") " pod="openstack/glance-30cc-account-create-5h92x" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.646604 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldc5s\" (UniqueName: \"kubernetes.io/projected/aeca1748-8fbf-4502-acb8-f1a2821a5c2b-kube-api-access-ldc5s\") pod \"placement-ca43-account-create-ts8dc\" (UID: \"aeca1748-8fbf-4502-acb8-f1a2821a5c2b\") " pod="openstack/placement-ca43-account-create-ts8dc" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.670858 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldc5s\" (UniqueName: \"kubernetes.io/projected/aeca1748-8fbf-4502-acb8-f1a2821a5c2b-kube-api-access-ldc5s\") pod \"placement-ca43-account-create-ts8dc\" (UID: \"aeca1748-8fbf-4502-acb8-f1a2821a5c2b\") " pod="openstack/placement-ca43-account-create-ts8dc" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.744812 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2824-account-create-j2pd4"] Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.748213 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jlkt\" (UniqueName: \"kubernetes.io/projected/00a1f32d-7cb9-403c-91aa-28f749dbbc9a-kube-api-access-2jlkt\") pod \"glance-30cc-account-create-5h92x\" (UID: \"00a1f32d-7cb9-403c-91aa-28f749dbbc9a\") " pod="openstack/glance-30cc-account-create-5h92x" Oct 10 09:27:36 crc kubenswrapper[4669]: W1010 09:27:36.752852 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd33626d7_d6dd_45c0_8442_1617295ec697.slice/crio-c67bac91eef58b3673ed67ec4ac6527c2f90a1cdb3d8682c822d70c3a6b6709a WatchSource:0}: Error finding container c67bac91eef58b3673ed67ec4ac6527c2f90a1cdb3d8682c822d70c3a6b6709a: Status 404 returned error can't find the container with id c67bac91eef58b3673ed67ec4ac6527c2f90a1cdb3d8682c822d70c3a6b6709a Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.776103 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca43-account-create-ts8dc" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.777068 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jlkt\" (UniqueName: \"kubernetes.io/projected/00a1f32d-7cb9-403c-91aa-28f749dbbc9a-kube-api-access-2jlkt\") pod \"glance-30cc-account-create-5h92x\" (UID: \"00a1f32d-7cb9-403c-91aa-28f749dbbc9a\") " pod="openstack/glance-30cc-account-create-5h92x" Oct 10 09:27:36 crc kubenswrapper[4669]: I1010 09:27:36.930853 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-30cc-account-create-5h92x" Oct 10 09:27:37 crc kubenswrapper[4669]: I1010 09:27:37.310297 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ca43-account-create-ts8dc"] Oct 10 09:27:37 crc kubenswrapper[4669]: I1010 09:27:37.402136 4669 generic.go:334] "Generic (PLEG): container finished" podID="d33626d7-d6dd-45c0-8442-1617295ec697" containerID="f909e76202798459913f1cf80d080b20b34f1b6de27025771dd02fabdf311693" exitCode=0 Oct 10 09:27:37 crc kubenswrapper[4669]: I1010 09:27:37.402214 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2824-account-create-j2pd4" event={"ID":"d33626d7-d6dd-45c0-8442-1617295ec697","Type":"ContainerDied","Data":"f909e76202798459913f1cf80d080b20b34f1b6de27025771dd02fabdf311693"} Oct 10 09:27:37 crc kubenswrapper[4669]: I1010 09:27:37.402242 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2824-account-create-j2pd4" event={"ID":"d33626d7-d6dd-45c0-8442-1617295ec697","Type":"ContainerStarted","Data":"c67bac91eef58b3673ed67ec4ac6527c2f90a1cdb3d8682c822d70c3a6b6709a"} Oct 10 09:27:37 crc kubenswrapper[4669]: I1010 09:27:37.403147 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ca43-account-create-ts8dc" event={"ID":"aeca1748-8fbf-4502-acb8-f1a2821a5c2b","Type":"ContainerStarted","Data":"6e6db5ce2348a43ed92fc0ca90f662bc3376070ecfc2b8d723c8ef836bee0309"} Oct 10 09:27:37 crc kubenswrapper[4669]: I1010 09:27:37.408138 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-30cc-account-create-5h92x"] Oct 10 09:27:37 crc kubenswrapper[4669]: I1010 09:27:37.817883 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.415384 4669 generic.go:334] "Generic (PLEG): container finished" podID="aeca1748-8fbf-4502-acb8-f1a2821a5c2b" containerID="462f97283198a175384ec8ae0d73744ccb89d9a7c2532829e0512bc21036aaff" exitCode=0 Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.415798 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ca43-account-create-ts8dc" event={"ID":"aeca1748-8fbf-4502-acb8-f1a2821a5c2b","Type":"ContainerDied","Data":"462f97283198a175384ec8ae0d73744ccb89d9a7c2532829e0512bc21036aaff"} Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.419227 4669 generic.go:334] "Generic (PLEG): container finished" podID="00a1f32d-7cb9-403c-91aa-28f749dbbc9a" containerID="0e5a72e7d38c9c30f1b85be8f13eb21e397b8e22f18dcd0d590b4ed94b10a110" exitCode=0 Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.419538 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-30cc-account-create-5h92x" event={"ID":"00a1f32d-7cb9-403c-91aa-28f749dbbc9a","Type":"ContainerDied","Data":"0e5a72e7d38c9c30f1b85be8f13eb21e397b8e22f18dcd0d590b4ed94b10a110"} Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.419559 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-30cc-account-create-5h92x" event={"ID":"00a1f32d-7cb9-403c-91aa-28f749dbbc9a","Type":"ContainerStarted","Data":"0a3a8aba29adcc07812616d30176e4a588e693d1bcbff828c7f35ac941796d91"} Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.731709 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2824-account-create-j2pd4" Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.784540 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfxm9\" (UniqueName: \"kubernetes.io/projected/d33626d7-d6dd-45c0-8442-1617295ec697-kube-api-access-tfxm9\") pod \"d33626d7-d6dd-45c0-8442-1617295ec697\" (UID: \"d33626d7-d6dd-45c0-8442-1617295ec697\") " Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.789977 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d33626d7-d6dd-45c0-8442-1617295ec697-kube-api-access-tfxm9" (OuterVolumeSpecName: "kube-api-access-tfxm9") pod "d33626d7-d6dd-45c0-8442-1617295ec697" (UID: "d33626d7-d6dd-45c0-8442-1617295ec697"). InnerVolumeSpecName "kube-api-access-tfxm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:38 crc kubenswrapper[4669]: I1010 09:27:38.886653 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfxm9\" (UniqueName: \"kubernetes.io/projected/d33626d7-d6dd-45c0-8442-1617295ec697-kube-api-access-tfxm9\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.430423 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2824-account-create-j2pd4" event={"ID":"d33626d7-d6dd-45c0-8442-1617295ec697","Type":"ContainerDied","Data":"c67bac91eef58b3673ed67ec4ac6527c2f90a1cdb3d8682c822d70c3a6b6709a"} Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.430490 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c67bac91eef58b3673ed67ec4ac6527c2f90a1cdb3d8682c822d70c3a6b6709a" Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.432251 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2824-account-create-j2pd4" Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.818342 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca43-account-create-ts8dc" Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.821845 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-30cc-account-create-5h92x" Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.905659 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jlkt\" (UniqueName: \"kubernetes.io/projected/00a1f32d-7cb9-403c-91aa-28f749dbbc9a-kube-api-access-2jlkt\") pod \"00a1f32d-7cb9-403c-91aa-28f749dbbc9a\" (UID: \"00a1f32d-7cb9-403c-91aa-28f749dbbc9a\") " Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.905753 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldc5s\" (UniqueName: \"kubernetes.io/projected/aeca1748-8fbf-4502-acb8-f1a2821a5c2b-kube-api-access-ldc5s\") pod \"aeca1748-8fbf-4502-acb8-f1a2821a5c2b\" (UID: \"aeca1748-8fbf-4502-acb8-f1a2821a5c2b\") " Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.910942 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeca1748-8fbf-4502-acb8-f1a2821a5c2b-kube-api-access-ldc5s" (OuterVolumeSpecName: "kube-api-access-ldc5s") pod "aeca1748-8fbf-4502-acb8-f1a2821a5c2b" (UID: "aeca1748-8fbf-4502-acb8-f1a2821a5c2b"). InnerVolumeSpecName "kube-api-access-ldc5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:39 crc kubenswrapper[4669]: I1010 09:27:39.911252 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a1f32d-7cb9-403c-91aa-28f749dbbc9a-kube-api-access-2jlkt" (OuterVolumeSpecName: "kube-api-access-2jlkt") pod "00a1f32d-7cb9-403c-91aa-28f749dbbc9a" (UID: "00a1f32d-7cb9-403c-91aa-28f749dbbc9a"). InnerVolumeSpecName "kube-api-access-2jlkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.007690 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jlkt\" (UniqueName: \"kubernetes.io/projected/00a1f32d-7cb9-403c-91aa-28f749dbbc9a-kube-api-access-2jlkt\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.007722 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldc5s\" (UniqueName: \"kubernetes.io/projected/aeca1748-8fbf-4502-acb8-f1a2821a5c2b-kube-api-access-ldc5s\") on node \"crc\" DevicePath \"\"" Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.440993 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-30cc-account-create-5h92x" Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.440973 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-30cc-account-create-5h92x" event={"ID":"00a1f32d-7cb9-403c-91aa-28f749dbbc9a","Type":"ContainerDied","Data":"0a3a8aba29adcc07812616d30176e4a588e693d1bcbff828c7f35ac941796d91"} Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.442289 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a3a8aba29adcc07812616d30176e4a588e693d1bcbff828c7f35ac941796d91" Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.442774 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ca43-account-create-ts8dc" event={"ID":"aeca1748-8fbf-4502-acb8-f1a2821a5c2b","Type":"ContainerDied","Data":"6e6db5ce2348a43ed92fc0ca90f662bc3376070ecfc2b8d723c8ef836bee0309"} Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.442805 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e6db5ce2348a43ed92fc0ca90f662bc3376070ecfc2b8d723c8ef836bee0309" Oct 10 09:27:40 crc kubenswrapper[4669]: I1010 09:27:40.442831 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca43-account-create-ts8dc" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.849620 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-99sxw"] Oct 10 09:27:41 crc kubenswrapper[4669]: E1010 09:27:41.850133 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeca1748-8fbf-4502-acb8-f1a2821a5c2b" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.850162 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeca1748-8fbf-4502-acb8-f1a2821a5c2b" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: E1010 09:27:41.850248 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33626d7-d6dd-45c0-8442-1617295ec697" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.850263 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33626d7-d6dd-45c0-8442-1617295ec697" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: E1010 09:27:41.850303 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a1f32d-7cb9-403c-91aa-28f749dbbc9a" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.850318 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a1f32d-7cb9-403c-91aa-28f749dbbc9a" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.850638 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a1f32d-7cb9-403c-91aa-28f749dbbc9a" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.850688 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d33626d7-d6dd-45c0-8442-1617295ec697" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.850726 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeca1748-8fbf-4502-acb8-f1a2821a5c2b" containerName="mariadb-account-create" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.851627 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.853700 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4lx7j" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.855701 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.860864 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-99sxw"] Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.938760 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-combined-ca-bundle\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.938806 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wws2h\" (UniqueName: \"kubernetes.io/projected/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-kube-api-access-wws2h\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.938871 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-db-sync-config-data\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:41 crc kubenswrapper[4669]: I1010 09:27:41.939022 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-config-data\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.040886 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-combined-ca-bundle\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.040962 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wws2h\" (UniqueName: \"kubernetes.io/projected/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-kube-api-access-wws2h\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.041030 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-db-sync-config-data\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.041095 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-config-data\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.047878 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-config-data\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.048247 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-db-sync-config-data\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.049011 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-combined-ca-bundle\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.066309 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wws2h\" (UniqueName: \"kubernetes.io/projected/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-kube-api-access-wws2h\") pod \"glance-db-sync-99sxw\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.189259 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-99sxw" Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.460696 4669 generic.go:334] "Generic (PLEG): container finished" podID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerID="971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3" exitCode=0 Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.460921 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b787ed5f-cbcf-4144-a87f-a014dbfd5282","Type":"ContainerDied","Data":"971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3"} Oct 10 09:27:42 crc kubenswrapper[4669]: I1010 09:27:42.754339 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-99sxw"] Oct 10 09:27:42 crc kubenswrapper[4669]: W1010 09:27:42.756955 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac28083f_b7bb_4f0e_aa1e_e1b69eaa8d53.slice/crio-d3a5bcb137324dc8cde20f82e8e52070ca13fc6733b0faa6910a2b09d953acc9 WatchSource:0}: Error finding container d3a5bcb137324dc8cde20f82e8e52070ca13fc6733b0faa6910a2b09d953acc9: Status 404 returned error can't find the container with id d3a5bcb137324dc8cde20f82e8e52070ca13fc6733b0faa6910a2b09d953acc9 Oct 10 09:27:43 crc kubenswrapper[4669]: I1010 09:27:43.487923 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-99sxw" event={"ID":"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53","Type":"ContainerStarted","Data":"d3a5bcb137324dc8cde20f82e8e52070ca13fc6733b0faa6910a2b09d953acc9"} Oct 10 09:27:43 crc kubenswrapper[4669]: I1010 09:27:43.493490 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b787ed5f-cbcf-4144-a87f-a014dbfd5282","Type":"ContainerStarted","Data":"e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80"} Oct 10 09:27:43 crc kubenswrapper[4669]: I1010 09:27:43.493756 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 10 09:27:43 crc kubenswrapper[4669]: I1010 09:27:43.497792 4669 generic.go:334] "Generic (PLEG): container finished" podID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerID="c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8" exitCode=0 Oct 10 09:27:43 crc kubenswrapper[4669]: I1010 09:27:43.497863 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c46f84af-548a-4390-8242-c48ff4fc25aa","Type":"ContainerDied","Data":"c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8"} Oct 10 09:27:43 crc kubenswrapper[4669]: I1010 09:27:43.543604 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.668667995 podStartE2EDuration="1m2.543564452s" podCreationTimestamp="2025-10-10 09:26:41 +0000 UTC" firstStartedPulling="2025-10-10 09:26:43.565387924 +0000 UTC m=+946.581406666" lastFinishedPulling="2025-10-10 09:27:04.440284381 +0000 UTC m=+967.456303123" observedRunningTime="2025-10-10 09:27:43.529091905 +0000 UTC m=+1006.545110657" watchObservedRunningTime="2025-10-10 09:27:43.543564452 +0000 UTC m=+1006.559583224" Oct 10 09:27:44 crc kubenswrapper[4669]: I1010 09:27:44.506434 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c46f84af-548a-4390-8242-c48ff4fc25aa","Type":"ContainerStarted","Data":"3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc"} Oct 10 09:27:44 crc kubenswrapper[4669]: I1010 09:27:44.507220 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:27:44 crc kubenswrapper[4669]: I1010 09:27:44.528829 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.672578231 podStartE2EDuration="1m3.528809813s" podCreationTimestamp="2025-10-10 09:26:41 +0000 UTC" firstStartedPulling="2025-10-10 09:26:43.167846667 +0000 UTC m=+946.183865409" lastFinishedPulling="2025-10-10 09:27:09.024078249 +0000 UTC m=+972.040096991" observedRunningTime="2025-10-10 09:27:44.525226898 +0000 UTC m=+1007.541245640" watchObservedRunningTime="2025-10-10 09:27:44.528809813 +0000 UTC m=+1007.544828555" Oct 10 09:27:45 crc kubenswrapper[4669]: I1010 09:27:45.674088 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jrw5q" podUID="2b815c31-96c7-47c3-b458-35945b55b326" containerName="ovn-controller" probeResult="failure" output=< Oct 10 09:27:45 crc kubenswrapper[4669]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 10 09:27:45 crc kubenswrapper[4669]: > Oct 10 09:27:50 crc kubenswrapper[4669]: I1010 09:27:50.677001 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jrw5q" podUID="2b815c31-96c7-47c3-b458-35945b55b326" containerName="ovn-controller" probeResult="failure" output=< Oct 10 09:27:50 crc kubenswrapper[4669]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 10 09:27:50 crc kubenswrapper[4669]: > Oct 10 09:27:50 crc kubenswrapper[4669]: I1010 09:27:50.747870 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:27:50 crc kubenswrapper[4669]: I1010 09:27:50.775812 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-d8wg9" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.009164 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jrw5q-config-qhv2l"] Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.013268 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.017248 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.029858 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jrw5q-config-qhv2l"] Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.192227 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run-ovn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.192280 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6hmn\" (UniqueName: \"kubernetes.io/projected/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-kube-api-access-x6hmn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.192372 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-log-ovn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.192392 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-scripts\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.192540 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-additional-scripts\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.192606 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294131 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-log-ovn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294171 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-scripts\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294208 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-additional-scripts\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294224 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294271 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run-ovn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294296 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6hmn\" (UniqueName: \"kubernetes.io/projected/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-kube-api-access-x6hmn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294505 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run-ovn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294536 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294596 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-log-ovn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.294885 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-additional-scripts\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.297323 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-scripts\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.328703 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6hmn\" (UniqueName: \"kubernetes.io/projected/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-kube-api-access-x6hmn\") pod \"ovn-controller-jrw5q-config-qhv2l\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:51 crc kubenswrapper[4669]: I1010 09:27:51.333967 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.034820 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.444671 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8nbn8"] Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.445647 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nbn8" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.477691 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8nbn8"] Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.535713 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-hxt77"] Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.536667 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hxt77" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.557011 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hxt77"] Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.640359 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klz7k\" (UniqueName: \"kubernetes.io/projected/964cd051-6afe-4834-a7b8-ccbe2c547f4e-kube-api-access-klz7k\") pod \"cinder-db-create-hxt77\" (UID: \"964cd051-6afe-4834-a7b8-ccbe2c547f4e\") " pod="openstack/cinder-db-create-hxt77" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.640608 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql6gb\" (UniqueName: \"kubernetes.io/projected/93c88504-7493-485a-a2ed-e0e62fb4543f-kube-api-access-ql6gb\") pod \"barbican-db-create-8nbn8\" (UID: \"93c88504-7493-485a-a2ed-e0e62fb4543f\") " pod="openstack/barbican-db-create-8nbn8" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.726062 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-tkg49"] Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.727097 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tkg49" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.740044 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tkg49"] Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.742195 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klz7k\" (UniqueName: \"kubernetes.io/projected/964cd051-6afe-4834-a7b8-ccbe2c547f4e-kube-api-access-klz7k\") pod \"cinder-db-create-hxt77\" (UID: \"964cd051-6afe-4834-a7b8-ccbe2c547f4e\") " pod="openstack/cinder-db-create-hxt77" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.742298 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql6gb\" (UniqueName: \"kubernetes.io/projected/93c88504-7493-485a-a2ed-e0e62fb4543f-kube-api-access-ql6gb\") pod \"barbican-db-create-8nbn8\" (UID: \"93c88504-7493-485a-a2ed-e0e62fb4543f\") " pod="openstack/barbican-db-create-8nbn8" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.794503 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klz7k\" (UniqueName: \"kubernetes.io/projected/964cd051-6afe-4834-a7b8-ccbe2c547f4e-kube-api-access-klz7k\") pod \"cinder-db-create-hxt77\" (UID: \"964cd051-6afe-4834-a7b8-ccbe2c547f4e\") " pod="openstack/cinder-db-create-hxt77" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.797848 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql6gb\" (UniqueName: \"kubernetes.io/projected/93c88504-7493-485a-a2ed-e0e62fb4543f-kube-api-access-ql6gb\") pod \"barbican-db-create-8nbn8\" (UID: \"93c88504-7493-485a-a2ed-e0e62fb4543f\") " pod="openstack/barbican-db-create-8nbn8" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.845438 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr6f4\" (UniqueName: \"kubernetes.io/projected/ddb14f43-b1a9-4ba3-b6db-734948a9c722-kube-api-access-rr6f4\") pod \"neutron-db-create-tkg49\" (UID: \"ddb14f43-b1a9-4ba3-b6db-734948a9c722\") " pod="openstack/neutron-db-create-tkg49" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.854702 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hxt77" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.947404 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr6f4\" (UniqueName: \"kubernetes.io/projected/ddb14f43-b1a9-4ba3-b6db-734948a9c722-kube-api-access-rr6f4\") pod \"neutron-db-create-tkg49\" (UID: \"ddb14f43-b1a9-4ba3-b6db-734948a9c722\") " pod="openstack/neutron-db-create-tkg49" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.989794 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-29prl"] Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.990796 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.997406 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r86r4" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.997411 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 09:27:53 crc kubenswrapper[4669]: I1010 09:27:53.997410 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.000405 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-29prl"] Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.003092 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.030129 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr6f4\" (UniqueName: \"kubernetes.io/projected/ddb14f43-b1a9-4ba3-b6db-734948a9c722-kube-api-access-rr6f4\") pod \"neutron-db-create-tkg49\" (UID: \"ddb14f43-b1a9-4ba3-b6db-734948a9c722\") " pod="openstack/neutron-db-create-tkg49" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.055667 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tkg49" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.056467 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-config-data\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.056519 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-combined-ca-bundle\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.056537 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssxb7\" (UniqueName: \"kubernetes.io/projected/680da54f-bdac-4014-8fd5-92f1cb5e715c-kube-api-access-ssxb7\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.060466 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nbn8" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.157223 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-config-data\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.157293 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-combined-ca-bundle\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.157317 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssxb7\" (UniqueName: \"kubernetes.io/projected/680da54f-bdac-4014-8fd5-92f1cb5e715c-kube-api-access-ssxb7\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.160591 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-config-data\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.160636 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-combined-ca-bundle\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.177814 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssxb7\" (UniqueName: \"kubernetes.io/projected/680da54f-bdac-4014-8fd5-92f1cb5e715c-kube-api-access-ssxb7\") pod \"keystone-db-sync-29prl\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:54 crc kubenswrapper[4669]: I1010 09:27:54.323421 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29prl" Oct 10 09:27:55 crc kubenswrapper[4669]: I1010 09:27:55.682105 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jrw5q" podUID="2b815c31-96c7-47c3-b458-35945b55b326" containerName="ovn-controller" probeResult="failure" output=< Oct 10 09:27:55 crc kubenswrapper[4669]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 10 09:27:55 crc kubenswrapper[4669]: > Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.498755 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hxt77"] Oct 10 09:27:56 crc kubenswrapper[4669]: W1010 09:27:56.515517 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod964cd051_6afe_4834_a7b8_ccbe2c547f4e.slice/crio-6f61d8250ec37bcdde678a4679d52f4c8a3bc76ec74da0a287b283e08e9ffda1 WatchSource:0}: Error finding container 6f61d8250ec37bcdde678a4679d52f4c8a3bc76ec74da0a287b283e08e9ffda1: Status 404 returned error can't find the container with id 6f61d8250ec37bcdde678a4679d52f4c8a3bc76ec74da0a287b283e08e9ffda1 Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.598507 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-29prl"] Oct 10 09:27:56 crc kubenswrapper[4669]: W1010 09:27:56.604748 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod680da54f_bdac_4014_8fd5_92f1cb5e715c.slice/crio-42e3338f1d5d7e4e7fad9fb416a3cb2e55c567a5253c8eb8bfb0bff610aa45dc WatchSource:0}: Error finding container 42e3338f1d5d7e4e7fad9fb416a3cb2e55c567a5253c8eb8bfb0bff610aa45dc: Status 404 returned error can't find the container with id 42e3338f1d5d7e4e7fad9fb416a3cb2e55c567a5253c8eb8bfb0bff610aa45dc Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.605234 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jrw5q-config-qhv2l"] Oct 10 09:27:56 crc kubenswrapper[4669]: W1010 09:27:56.607598 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9e337d4_a3a0_44a3_a72d_64ca4fbd760e.slice/crio-dc917e8ba76a504579a9b099830270d32969a8c01befd538042bd1cde167a4a1 WatchSource:0}: Error finding container dc917e8ba76a504579a9b099830270d32969a8c01befd538042bd1cde167a4a1: Status 404 returned error can't find the container with id dc917e8ba76a504579a9b099830270d32969a8c01befd538042bd1cde167a4a1 Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.637715 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hxt77" event={"ID":"964cd051-6afe-4834-a7b8-ccbe2c547f4e","Type":"ContainerStarted","Data":"6f61d8250ec37bcdde678a4679d52f4c8a3bc76ec74da0a287b283e08e9ffda1"} Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.638965 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29prl" event={"ID":"680da54f-bdac-4014-8fd5-92f1cb5e715c","Type":"ContainerStarted","Data":"42e3338f1d5d7e4e7fad9fb416a3cb2e55c567a5253c8eb8bfb0bff610aa45dc"} Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.640116 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q-config-qhv2l" event={"ID":"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e","Type":"ContainerStarted","Data":"dc917e8ba76a504579a9b099830270d32969a8c01befd538042bd1cde167a4a1"} Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.747749 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tkg49"] Oct 10 09:27:56 crc kubenswrapper[4669]: I1010 09:27:56.756066 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8nbn8"] Oct 10 09:27:56 crc kubenswrapper[4669]: W1010 09:27:56.769736 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddb14f43_b1a9_4ba3_b6db_734948a9c722.slice/crio-1db7a21d3b49b4f8ea557510dab0da720c2c7982c801ee729ff5c38de7c35ef4 WatchSource:0}: Error finding container 1db7a21d3b49b4f8ea557510dab0da720c2c7982c801ee729ff5c38de7c35ef4: Status 404 returned error can't find the container with id 1db7a21d3b49b4f8ea557510dab0da720c2c7982c801ee729ff5c38de7c35ef4 Oct 10 09:27:56 crc kubenswrapper[4669]: W1010 09:27:56.770228 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93c88504_7493_485a_a2ed_e0e62fb4543f.slice/crio-7aa2f7c3a815325230782dc3cac763a012e84b64330b08ad31ca20ca9340cb03 WatchSource:0}: Error finding container 7aa2f7c3a815325230782dc3cac763a012e84b64330b08ad31ca20ca9340cb03: Status 404 returned error can't find the container with id 7aa2f7c3a815325230782dc3cac763a012e84b64330b08ad31ca20ca9340cb03 Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.661918 4669 generic.go:334] "Generic (PLEG): container finished" podID="964cd051-6afe-4834-a7b8-ccbe2c547f4e" containerID="c38c3547c8ddcc5a50d07b3f2d76d69c047123990c58e1bb914f36b792f41196" exitCode=0 Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.662110 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hxt77" event={"ID":"964cd051-6afe-4834-a7b8-ccbe2c547f4e","Type":"ContainerDied","Data":"c38c3547c8ddcc5a50d07b3f2d76d69c047123990c58e1bb914f36b792f41196"} Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.668830 4669 generic.go:334] "Generic (PLEG): container finished" podID="93c88504-7493-485a-a2ed-e0e62fb4543f" containerID="c418e67ac115abd378150868c2caee1b29bf9cfa73128fa8758b011bed59ad52" exitCode=0 Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.668896 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8nbn8" event={"ID":"93c88504-7493-485a-a2ed-e0e62fb4543f","Type":"ContainerDied","Data":"c418e67ac115abd378150868c2caee1b29bf9cfa73128fa8758b011bed59ad52"} Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.668917 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8nbn8" event={"ID":"93c88504-7493-485a-a2ed-e0e62fb4543f","Type":"ContainerStarted","Data":"7aa2f7c3a815325230782dc3cac763a012e84b64330b08ad31ca20ca9340cb03"} Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.670953 4669 generic.go:334] "Generic (PLEG): container finished" podID="b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" containerID="5a021356d80a083e165c66ee1fb989416744bf9171ad4471c6138eb28958ca9a" exitCode=0 Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.671044 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q-config-qhv2l" event={"ID":"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e","Type":"ContainerDied","Data":"5a021356d80a083e165c66ee1fb989416744bf9171ad4471c6138eb28958ca9a"} Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.674052 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-99sxw" event={"ID":"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53","Type":"ContainerStarted","Data":"51543363f25e9768823f820b2bbaadd13707ccd3a1803e5b0408d0d5c003ecdc"} Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.681869 4669 generic.go:334] "Generic (PLEG): container finished" podID="ddb14f43-b1a9-4ba3-b6db-734948a9c722" containerID="d2168a9083391f5a4c7edbf68ea3c7b9810c20f575156eca65bae1c65db4e44c" exitCode=0 Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.681954 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tkg49" event={"ID":"ddb14f43-b1a9-4ba3-b6db-734948a9c722","Type":"ContainerDied","Data":"d2168a9083391f5a4c7edbf68ea3c7b9810c20f575156eca65bae1c65db4e44c"} Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.682085 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tkg49" event={"ID":"ddb14f43-b1a9-4ba3-b6db-734948a9c722","Type":"ContainerStarted","Data":"1db7a21d3b49b4f8ea557510dab0da720c2c7982c801ee729ff5c38de7c35ef4"} Oct 10 09:27:57 crc kubenswrapper[4669]: I1010 09:27:57.718244 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-99sxw" podStartSLOduration=3.438089675 podStartE2EDuration="16.718225509s" podCreationTimestamp="2025-10-10 09:27:41 +0000 UTC" firstStartedPulling="2025-10-10 09:27:42.758935706 +0000 UTC m=+1005.774954448" lastFinishedPulling="2025-10-10 09:27:56.03907153 +0000 UTC m=+1019.055090282" observedRunningTime="2025-10-10 09:27:57.714674555 +0000 UTC m=+1020.730693297" watchObservedRunningTime="2025-10-10 09:27:57.718225509 +0000 UTC m=+1020.734244251" Oct 10 09:28:00 crc kubenswrapper[4669]: I1010 09:28:00.687966 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jrw5q" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.333129 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nbn8" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.346111 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.475876 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-scripts\") pod \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.475917 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6hmn\" (UniqueName: \"kubernetes.io/projected/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-kube-api-access-x6hmn\") pod \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476005 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-additional-scripts\") pod \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476035 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql6gb\" (UniqueName: \"kubernetes.io/projected/93c88504-7493-485a-a2ed-e0e62fb4543f-kube-api-access-ql6gb\") pod \"93c88504-7493-485a-a2ed-e0e62fb4543f\" (UID: \"93c88504-7493-485a-a2ed-e0e62fb4543f\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476074 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-log-ovn\") pod \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476090 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run\") pod \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476120 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run-ovn\") pod \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\" (UID: \"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476415 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" (UID: "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476446 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" (UID: "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.476460 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run" (OuterVolumeSpecName: "var-run") pod "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" (UID: "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.477007 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-scripts" (OuterVolumeSpecName: "scripts") pod "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" (UID: "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.477209 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" (UID: "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.479570 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-kube-api-access-x6hmn" (OuterVolumeSpecName: "kube-api-access-x6hmn") pod "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" (UID: "b9e337d4-a3a0-44a3-a72d-64ca4fbd760e"). InnerVolumeSpecName "kube-api-access-x6hmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.481874 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c88504-7493-485a-a2ed-e0e62fb4543f-kube-api-access-ql6gb" (OuterVolumeSpecName: "kube-api-access-ql6gb") pod "93c88504-7493-485a-a2ed-e0e62fb4543f" (UID: "93c88504-7493-485a-a2ed-e0e62fb4543f"). InnerVolumeSpecName "kube-api-access-ql6gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.512725 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hxt77" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.525467 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tkg49" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.577929 4669 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.577964 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql6gb\" (UniqueName: \"kubernetes.io/projected/93c88504-7493-485a-a2ed-e0e62fb4543f-kube-api-access-ql6gb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.577976 4669 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.577985 4669 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.577994 4669 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.578003 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.578010 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6hmn\" (UniqueName: \"kubernetes.io/projected/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e-kube-api-access-x6hmn\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.679686 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klz7k\" (UniqueName: \"kubernetes.io/projected/964cd051-6afe-4834-a7b8-ccbe2c547f4e-kube-api-access-klz7k\") pod \"964cd051-6afe-4834-a7b8-ccbe2c547f4e\" (UID: \"964cd051-6afe-4834-a7b8-ccbe2c547f4e\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.679813 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr6f4\" (UniqueName: \"kubernetes.io/projected/ddb14f43-b1a9-4ba3-b6db-734948a9c722-kube-api-access-rr6f4\") pod \"ddb14f43-b1a9-4ba3-b6db-734948a9c722\" (UID: \"ddb14f43-b1a9-4ba3-b6db-734948a9c722\") " Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.689665 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964cd051-6afe-4834-a7b8-ccbe2c547f4e-kube-api-access-klz7k" (OuterVolumeSpecName: "kube-api-access-klz7k") pod "964cd051-6afe-4834-a7b8-ccbe2c547f4e" (UID: "964cd051-6afe-4834-a7b8-ccbe2c547f4e"). InnerVolumeSpecName "kube-api-access-klz7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.691345 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddb14f43-b1a9-4ba3-b6db-734948a9c722-kube-api-access-rr6f4" (OuterVolumeSpecName: "kube-api-access-rr6f4") pod "ddb14f43-b1a9-4ba3-b6db-734948a9c722" (UID: "ddb14f43-b1a9-4ba3-b6db-734948a9c722"). InnerVolumeSpecName "kube-api-access-rr6f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.719976 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8nbn8" event={"ID":"93c88504-7493-485a-a2ed-e0e62fb4543f","Type":"ContainerDied","Data":"7aa2f7c3a815325230782dc3cac763a012e84b64330b08ad31ca20ca9340cb03"} Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.720002 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8nbn8" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.720019 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7aa2f7c3a815325230782dc3cac763a012e84b64330b08ad31ca20ca9340cb03" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.721634 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29prl" event={"ID":"680da54f-bdac-4014-8fd5-92f1cb5e715c","Type":"ContainerStarted","Data":"c2b05e56b87acff4dc70c091af8c816c8013320e03be0c03e899054fbb519492"} Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.724543 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-qhv2l" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.724799 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q-config-qhv2l" event={"ID":"b9e337d4-a3a0-44a3-a72d-64ca4fbd760e","Type":"ContainerDied","Data":"dc917e8ba76a504579a9b099830270d32969a8c01befd538042bd1cde167a4a1"} Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.724886 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc917e8ba76a504579a9b099830270d32969a8c01befd538042bd1cde167a4a1" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.728169 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tkg49" event={"ID":"ddb14f43-b1a9-4ba3-b6db-734948a9c722","Type":"ContainerDied","Data":"1db7a21d3b49b4f8ea557510dab0da720c2c7982c801ee729ff5c38de7c35ef4"} Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.728203 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1db7a21d3b49b4f8ea557510dab0da720c2c7982c801ee729ff5c38de7c35ef4" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.728276 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tkg49" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.730287 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hxt77" event={"ID":"964cd051-6afe-4834-a7b8-ccbe2c547f4e","Type":"ContainerDied","Data":"6f61d8250ec37bcdde678a4679d52f4c8a3bc76ec74da0a287b283e08e9ffda1"} Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.730340 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f61d8250ec37bcdde678a4679d52f4c8a3bc76ec74da0a287b283e08e9ffda1" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.730405 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hxt77" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.745115 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-29prl" podStartSLOduration=4.019911959 podStartE2EDuration="8.745095093s" podCreationTimestamp="2025-10-10 09:27:53 +0000 UTC" firstStartedPulling="2025-10-10 09:27:56.610510778 +0000 UTC m=+1019.626529520" lastFinishedPulling="2025-10-10 09:28:01.335693912 +0000 UTC m=+1024.351712654" observedRunningTime="2025-10-10 09:28:01.742283232 +0000 UTC m=+1024.758301974" watchObservedRunningTime="2025-10-10 09:28:01.745095093 +0000 UTC m=+1024.761113825" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.783281 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klz7k\" (UniqueName: \"kubernetes.io/projected/964cd051-6afe-4834-a7b8-ccbe2c547f4e-kube-api-access-klz7k\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: I1010 09:28:01.783327 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr6f4\" (UniqueName: \"kubernetes.io/projected/ddb14f43-b1a9-4ba3-b6db-734948a9c722-kube-api-access-rr6f4\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:01 crc kubenswrapper[4669]: E1010 09:28:01.901166 4669 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93c88504_7493_485a_a2ed_e0e62fb4543f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddb14f43_b1a9_4ba3_b6db_734948a9c722.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod964cd051_6afe_4834_a7b8_ccbe2c547f4e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93c88504_7493_485a_a2ed_e0e62fb4543f.slice/crio-7aa2f7c3a815325230782dc3cac763a012e84b64330b08ad31ca20ca9340cb03\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod964cd051_6afe_4834_a7b8_ccbe2c547f4e.slice/crio-6f61d8250ec37bcdde678a4679d52f4c8a3bc76ec74da0a287b283e08e9ffda1\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9e337d4_a3a0_44a3_a72d_64ca4fbd760e.slice\": RecentStats: unable to find data in memory cache]" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.467285 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jrw5q-config-qhv2l"] Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.474940 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jrw5q-config-qhv2l"] Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.515621 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.603187 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jrw5q-config-2qsfj"] Oct 10 09:28:02 crc kubenswrapper[4669]: E1010 09:28:02.603785 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964cd051-6afe-4834-a7b8-ccbe2c547f4e" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.603869 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="964cd051-6afe-4834-a7b8-ccbe2c547f4e" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: E1010 09:28:02.603944 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c88504-7493-485a-a2ed-e0e62fb4543f" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.604002 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c88504-7493-485a-a2ed-e0e62fb4543f" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: E1010 09:28:02.604061 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" containerName="ovn-config" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.604120 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" containerName="ovn-config" Oct 10 09:28:02 crc kubenswrapper[4669]: E1010 09:28:02.604182 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb14f43-b1a9-4ba3-b6db-734948a9c722" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.604242 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb14f43-b1a9-4ba3-b6db-734948a9c722" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.604441 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" containerName="ovn-config" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.604514 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="964cd051-6afe-4834-a7b8-ccbe2c547f4e" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.604574 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddb14f43-b1a9-4ba3-b6db-734948a9c722" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.604684 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c88504-7493-485a-a2ed-e0e62fb4543f" containerName="mariadb-database-create" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.605274 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.608234 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.627531 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jrw5q-config-2qsfj"] Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.799945 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-log-ovn\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.799991 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-additional-scripts\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.800020 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd9rq\" (UniqueName: \"kubernetes.io/projected/47234f06-29c4-428d-91aa-101dd8f2e806-kube-api-access-hd9rq\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.800121 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.800172 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-scripts\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.800252 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run-ovn\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.902453 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.902773 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.902874 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-scripts\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.903076 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run-ovn\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.903143 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run-ovn\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.903358 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-log-ovn\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.903411 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-additional-scripts\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.903455 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd9rq\" (UniqueName: \"kubernetes.io/projected/47234f06-29c4-428d-91aa-101dd8f2e806-kube-api-access-hd9rq\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.903499 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-log-ovn\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.928000 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-additional-scripts\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.930857 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-scripts\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:02 crc kubenswrapper[4669]: I1010 09:28:02.932261 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd9rq\" (UniqueName: \"kubernetes.io/projected/47234f06-29c4-428d-91aa-101dd8f2e806-kube-api-access-hd9rq\") pod \"ovn-controller-jrw5q-config-2qsfj\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:03 crc kubenswrapper[4669]: I1010 09:28:03.229173 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:03 crc kubenswrapper[4669]: I1010 09:28:03.728904 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jrw5q-config-2qsfj"] Oct 10 09:28:03 crc kubenswrapper[4669]: I1010 09:28:03.752639 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q-config-2qsfj" event={"ID":"47234f06-29c4-428d-91aa-101dd8f2e806","Type":"ContainerStarted","Data":"23a13da709f5b932789b515dccdc5d00d7d0bb5138ef38f5d165c933fa022acb"} Oct 10 09:28:03 crc kubenswrapper[4669]: I1010 09:28:03.807699 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9e337d4-a3a0-44a3-a72d-64ca4fbd760e" path="/var/lib/kubelet/pods/b9e337d4-a3a0-44a3-a72d-64ca4fbd760e/volumes" Oct 10 09:28:04 crc kubenswrapper[4669]: I1010 09:28:04.781659 4669 generic.go:334] "Generic (PLEG): container finished" podID="ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" containerID="51543363f25e9768823f820b2bbaadd13707ccd3a1803e5b0408d0d5c003ecdc" exitCode=0 Oct 10 09:28:04 crc kubenswrapper[4669]: I1010 09:28:04.781734 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-99sxw" event={"ID":"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53","Type":"ContainerDied","Data":"51543363f25e9768823f820b2bbaadd13707ccd3a1803e5b0408d0d5c003ecdc"} Oct 10 09:28:04 crc kubenswrapper[4669]: I1010 09:28:04.786977 4669 generic.go:334] "Generic (PLEG): container finished" podID="47234f06-29c4-428d-91aa-101dd8f2e806" containerID="d37d3fd44e8c91fade589abd164824cb7e60a615a7ec70be22037907b2ee5cb6" exitCode=0 Oct 10 09:28:04 crc kubenswrapper[4669]: I1010 09:28:04.787041 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q-config-2qsfj" event={"ID":"47234f06-29c4-428d-91aa-101dd8f2e806","Type":"ContainerDied","Data":"d37d3fd44e8c91fade589abd164824cb7e60a615a7ec70be22037907b2ee5cb6"} Oct 10 09:28:05 crc kubenswrapper[4669]: I1010 09:28:05.797388 4669 generic.go:334] "Generic (PLEG): container finished" podID="680da54f-bdac-4014-8fd5-92f1cb5e715c" containerID="c2b05e56b87acff4dc70c091af8c816c8013320e03be0c03e899054fbb519492" exitCode=0 Oct 10 09:28:05 crc kubenswrapper[4669]: I1010 09:28:05.819430 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29prl" event={"ID":"680da54f-bdac-4014-8fd5-92f1cb5e715c","Type":"ContainerDied","Data":"c2b05e56b87acff4dc70c091af8c816c8013320e03be0c03e899054fbb519492"} Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.313344 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.320284 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-99sxw" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.464256 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-db-sync-config-data\") pod \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.464904 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run\") pod \"47234f06-29c4-428d-91aa-101dd8f2e806\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465000 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-log-ovn\") pod \"47234f06-29c4-428d-91aa-101dd8f2e806\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465106 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wws2h\" (UniqueName: \"kubernetes.io/projected/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-kube-api-access-wws2h\") pod \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465195 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-config-data\") pod \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465333 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-scripts\") pod \"47234f06-29c4-428d-91aa-101dd8f2e806\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465467 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd9rq\" (UniqueName: \"kubernetes.io/projected/47234f06-29c4-428d-91aa-101dd8f2e806-kube-api-access-hd9rq\") pod \"47234f06-29c4-428d-91aa-101dd8f2e806\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465563 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-additional-scripts\") pod \"47234f06-29c4-428d-91aa-101dd8f2e806\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465763 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-combined-ca-bundle\") pod \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\" (UID: \"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465899 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run-ovn\") pod \"47234f06-29c4-428d-91aa-101dd8f2e806\" (UID: \"47234f06-29c4-428d-91aa-101dd8f2e806\") " Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465002 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run" (OuterVolumeSpecName: "var-run") pod "47234f06-29c4-428d-91aa-101dd8f2e806" (UID: "47234f06-29c4-428d-91aa-101dd8f2e806"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.465025 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "47234f06-29c4-428d-91aa-101dd8f2e806" (UID: "47234f06-29c4-428d-91aa-101dd8f2e806"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.466233 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-scripts" (OuterVolumeSpecName: "scripts") pod "47234f06-29c4-428d-91aa-101dd8f2e806" (UID: "47234f06-29c4-428d-91aa-101dd8f2e806"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.466253 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "47234f06-29c4-428d-91aa-101dd8f2e806" (UID: "47234f06-29c4-428d-91aa-101dd8f2e806"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.466456 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "47234f06-29c4-428d-91aa-101dd8f2e806" (UID: "47234f06-29c4-428d-91aa-101dd8f2e806"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.470235 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-kube-api-access-wws2h" (OuterVolumeSpecName: "kube-api-access-wws2h") pod "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" (UID: "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53"). InnerVolumeSpecName "kube-api-access-wws2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.470828 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" (UID: "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.482739 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47234f06-29c4-428d-91aa-101dd8f2e806-kube-api-access-hd9rq" (OuterVolumeSpecName: "kube-api-access-hd9rq") pod "47234f06-29c4-428d-91aa-101dd8f2e806" (UID: "47234f06-29c4-428d-91aa-101dd8f2e806"). InnerVolumeSpecName "kube-api-access-hd9rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.489195 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" (UID: "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.509050 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-config-data" (OuterVolumeSpecName: "config-data") pod "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" (UID: "ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568006 4669 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568037 4669 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568048 4669 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-run\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568056 4669 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47234f06-29c4-428d-91aa-101dd8f2e806-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568064 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wws2h\" (UniqueName: \"kubernetes.io/projected/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-kube-api-access-wws2h\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568072 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568083 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568095 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd9rq\" (UniqueName: \"kubernetes.io/projected/47234f06-29c4-428d-91aa-101dd8f2e806-kube-api-access-hd9rq\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568110 4669 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/47234f06-29c4-428d-91aa-101dd8f2e806-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.568118 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.808370 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-99sxw" event={"ID":"ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53","Type":"ContainerDied","Data":"d3a5bcb137324dc8cde20f82e8e52070ca13fc6733b0faa6910a2b09d953acc9"} Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.808412 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3a5bcb137324dc8cde20f82e8e52070ca13fc6733b0faa6910a2b09d953acc9" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.808389 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-99sxw" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.811372 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jrw5q-config-2qsfj" event={"ID":"47234f06-29c4-428d-91aa-101dd8f2e806","Type":"ContainerDied","Data":"23a13da709f5b932789b515dccdc5d00d7d0bb5138ef38f5d165c933fa022acb"} Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.811434 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23a13da709f5b932789b515dccdc5d00d7d0bb5138ef38f5d165c933fa022acb" Oct 10 09:28:06 crc kubenswrapper[4669]: I1010 09:28:06.811500 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jrw5q-config-2qsfj" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.089215 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29prl" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.180122 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-combined-ca-bundle\") pod \"680da54f-bdac-4014-8fd5-92f1cb5e715c\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.180211 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssxb7\" (UniqueName: \"kubernetes.io/projected/680da54f-bdac-4014-8fd5-92f1cb5e715c-kube-api-access-ssxb7\") pod \"680da54f-bdac-4014-8fd5-92f1cb5e715c\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.180246 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-config-data\") pod \"680da54f-bdac-4014-8fd5-92f1cb5e715c\" (UID: \"680da54f-bdac-4014-8fd5-92f1cb5e715c\") " Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.188525 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/680da54f-bdac-4014-8fd5-92f1cb5e715c-kube-api-access-ssxb7" (OuterVolumeSpecName: "kube-api-access-ssxb7") pod "680da54f-bdac-4014-8fd5-92f1cb5e715c" (UID: "680da54f-bdac-4014-8fd5-92f1cb5e715c"). InnerVolumeSpecName "kube-api-access-ssxb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.269875 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "680da54f-bdac-4014-8fd5-92f1cb5e715c" (UID: "680da54f-bdac-4014-8fd5-92f1cb5e715c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.282265 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.282294 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssxb7\" (UniqueName: \"kubernetes.io/projected/680da54f-bdac-4014-8fd5-92f1cb5e715c-kube-api-access-ssxb7\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.316398 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-config-data" (OuterVolumeSpecName: "config-data") pod "680da54f-bdac-4014-8fd5-92f1cb5e715c" (UID: "680da54f-bdac-4014-8fd5-92f1cb5e715c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.383669 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/680da54f-bdac-4014-8fd5-92f1cb5e715c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.435922 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jrw5q-config-2qsfj"] Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.443398 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jrw5q-config-2qsfj"] Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.501689 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bspk7"] Oct 10 09:28:07 crc kubenswrapper[4669]: E1010 09:28:07.502055 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680da54f-bdac-4014-8fd5-92f1cb5e715c" containerName="keystone-db-sync" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.502077 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="680da54f-bdac-4014-8fd5-92f1cb5e715c" containerName="keystone-db-sync" Oct 10 09:28:07 crc kubenswrapper[4669]: E1010 09:28:07.502092 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47234f06-29c4-428d-91aa-101dd8f2e806" containerName="ovn-config" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.502100 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47234f06-29c4-428d-91aa-101dd8f2e806" containerName="ovn-config" Oct 10 09:28:07 crc kubenswrapper[4669]: E1010 09:28:07.502116 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" containerName="glance-db-sync" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.502123 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" containerName="glance-db-sync" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.502329 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" containerName="glance-db-sync" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.502346 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="47234f06-29c4-428d-91aa-101dd8f2e806" containerName="ovn-config" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.502357 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="680da54f-bdac-4014-8fd5-92f1cb5e715c" containerName="keystone-db-sync" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.503295 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.559381 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bspk7"] Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.586879 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-sb\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.586925 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-dns-svc\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.586976 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-nb\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.587035 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb4wd\" (UniqueName: \"kubernetes.io/projected/7428b8ac-84e1-4438-944f-e98f085f8055-kube-api-access-sb4wd\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.587074 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-config\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.688048 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb4wd\" (UniqueName: \"kubernetes.io/projected/7428b8ac-84e1-4438-944f-e98f085f8055-kube-api-access-sb4wd\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.688110 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-config\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.688151 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-sb\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.688188 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-dns-svc\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.689176 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-sb\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.689240 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-nb\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.689294 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-config\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.689344 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-dns-svc\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.689794 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-nb\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.703653 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb4wd\" (UniqueName: \"kubernetes.io/projected/7428b8ac-84e1-4438-944f-e98f085f8055-kube-api-access-sb4wd\") pod \"dnsmasq-dns-74b7749bc7-bspk7\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.807639 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47234f06-29c4-428d-91aa-101dd8f2e806" path="/var/lib/kubelet/pods/47234f06-29c4-428d-91aa-101dd8f2e806/volumes" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.825101 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.859418 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-29prl" event={"ID":"680da54f-bdac-4014-8fd5-92f1cb5e715c","Type":"ContainerDied","Data":"42e3338f1d5d7e4e7fad9fb416a3cb2e55c567a5253c8eb8bfb0bff610aa45dc"} Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.859459 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42e3338f1d5d7e4e7fad9fb416a3cb2e55c567a5253c8eb8bfb0bff610aa45dc" Oct 10 09:28:07 crc kubenswrapper[4669]: I1010 09:28:07.859534 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-29prl" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.106436 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bspk7"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.146130 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-4htn5"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.147431 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.160552 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-4htn5"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.212829 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hq8kp"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.213944 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.225040 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hq8kp"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.230543 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.230746 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.230855 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r86r4" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.231011 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 09:28:08 crc kubenswrapper[4669]: W1010 09:28:08.260768 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7428b8ac_84e1_4438_944f_e98f085f8055.slice/crio-538fc3fe9508bdfa851dd98df70a1b51225ca8ea77bca7928af92675efd19448 WatchSource:0}: Error finding container 538fc3fe9508bdfa851dd98df70a1b51225ca8ea77bca7928af92675efd19448: Status 404 returned error can't find the container with id 538fc3fe9508bdfa851dd98df70a1b51225ca8ea77bca7928af92675efd19448 Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.285165 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bspk7"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.299961 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-fernet-keys\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.300276 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb2bt\" (UniqueName: \"kubernetes.io/projected/139fea85-7c6e-45de-b9f9-d667c361a9d1-kube-api-access-tb2bt\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.300307 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-credential-keys\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.300335 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-config\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.300743 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-dns-svc\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.300833 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.300907 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.300984 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-combined-ca-bundle\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.301065 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t4st\" (UniqueName: \"kubernetes.io/projected/61c92eb8-effe-434f-9c71-cb74467723e9-kube-api-access-7t4st\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.301147 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-scripts\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.301230 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-config-data\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426151 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-config\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426278 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-dns-svc\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426318 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426346 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426369 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-combined-ca-bundle\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426399 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t4st\" (UniqueName: \"kubernetes.io/projected/61c92eb8-effe-434f-9c71-cb74467723e9-kube-api-access-7t4st\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426436 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-scripts\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426480 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-config-data\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426547 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-fernet-keys\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426565 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb2bt\" (UniqueName: \"kubernetes.io/projected/139fea85-7c6e-45de-b9f9-d667c361a9d1-kube-api-access-tb2bt\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.426611 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-credential-keys\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.431309 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-combined-ca-bundle\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.431979 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-config\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.432453 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-dns-svc\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.433137 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.433793 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.434188 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-credential-keys\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.438540 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-fernet-keys\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.439085 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-config-data\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.451890 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-scripts\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.458914 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb2bt\" (UniqueName: \"kubernetes.io/projected/139fea85-7c6e-45de-b9f9-d667c361a9d1-kube-api-access-tb2bt\") pod \"dnsmasq-dns-67bcfd764f-4htn5\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.477400 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t4st\" (UniqueName: \"kubernetes.io/projected/61c92eb8-effe-434f-9c71-cb74467723e9-kube-api-access-7t4st\") pod \"keystone-bootstrap-hq8kp\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.517806 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.519660 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.531813 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.533303 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.533524 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.555981 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.603744 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.634528 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-4htn5"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.635348 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-config-data\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.635372 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-log-httpd\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.635389 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-scripts\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.635424 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.635442 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9rtc\" (UniqueName: \"kubernetes.io/projected/7f64ef22-59f4-471b-9726-b3673ba3ddcf-kube-api-access-c9rtc\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.635511 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-run-httpd\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.635532 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.687571 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-mcjwv"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.702949 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.717607 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-mcjwv"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.724024 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-z8vqt"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.725716 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736116 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736406 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-8hzxw" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736398 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736493 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-config-data\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736542 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-log-httpd\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736563 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-scripts\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736604 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736619 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736639 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9rtc\" (UniqueName: \"kubernetes.io/projected/7f64ef22-59f4-471b-9726-b3673ba3ddcf-kube-api-access-c9rtc\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.736669 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-run-httpd\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.737008 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-run-httpd\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.737249 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-log-httpd\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.743005 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-scripts\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.746148 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.749403 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-config-data\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.751163 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.774048 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9rtc\" (UniqueName: \"kubernetes.io/projected/7f64ef22-59f4-471b-9726-b3673ba3ddcf-kube-api-access-c9rtc\") pod \"ceilometer-0\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.786122 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z8vqt"] Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838381 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-nb\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838442 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-combined-ca-bundle\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838472 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/799cf433-723c-4168-8522-773ca2947ec3-logs\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838492 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-scripts\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838521 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-config-data\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838538 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqx4b\" (UniqueName: \"kubernetes.io/projected/60f556bd-2a52-4976-8709-0d79392c793e-kube-api-access-jqx4b\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838618 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glkr2\" (UniqueName: \"kubernetes.io/projected/799cf433-723c-4168-8522-773ca2947ec3-kube-api-access-glkr2\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838665 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-dns-svc\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838692 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-config\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.838708 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-sb\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.856192 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.875380 4669 generic.go:334] "Generic (PLEG): container finished" podID="7428b8ac-84e1-4438-944f-e98f085f8055" containerID="4125334e23c9f137fe1abec077087863f6965bcbba1ff246fe71af2515f654c5" exitCode=0 Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.875732 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" event={"ID":"7428b8ac-84e1-4438-944f-e98f085f8055","Type":"ContainerDied","Data":"4125334e23c9f137fe1abec077087863f6965bcbba1ff246fe71af2515f654c5"} Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.875760 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" event={"ID":"7428b8ac-84e1-4438-944f-e98f085f8055","Type":"ContainerStarted","Data":"538fc3fe9508bdfa851dd98df70a1b51225ca8ea77bca7928af92675efd19448"} Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943420 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-nb\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943484 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-combined-ca-bundle\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943517 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/799cf433-723c-4168-8522-773ca2947ec3-logs\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943539 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-scripts\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943560 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-config-data\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943577 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqx4b\" (UniqueName: \"kubernetes.io/projected/60f556bd-2a52-4976-8709-0d79392c793e-kube-api-access-jqx4b\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943609 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glkr2\" (UniqueName: \"kubernetes.io/projected/799cf433-723c-4168-8522-773ca2947ec3-kube-api-access-glkr2\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943657 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-dns-svc\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943684 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-config\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.943706 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-sb\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.945831 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-nb\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.947253 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-config\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.947484 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-sb\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.948074 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-dns-svc\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.948385 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/799cf433-723c-4168-8522-773ca2947ec3-logs\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.949173 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-combined-ca-bundle\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.949385 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-scripts\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.950061 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-config-data\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.961325 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glkr2\" (UniqueName: \"kubernetes.io/projected/799cf433-723c-4168-8522-773ca2947ec3-kube-api-access-glkr2\") pod \"placement-db-sync-z8vqt\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:08 crc kubenswrapper[4669]: I1010 09:28:08.972543 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqx4b\" (UniqueName: \"kubernetes.io/projected/60f556bd-2a52-4976-8709-0d79392c793e-kube-api-access-jqx4b\") pod \"dnsmasq-dns-7b99bccc6c-mcjwv\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:09 crc kubenswrapper[4669]: W1010 09:28:09.029785 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61c92eb8_effe_434f_9c71_cb74467723e9.slice/crio-2a715aa0e7459de84446d58d08828b54953cb8242a142c7dc7d71b772f2879e8 WatchSource:0}: Error finding container 2a715aa0e7459de84446d58d08828b54953cb8242a142c7dc7d71b772f2879e8: Status 404 returned error can't find the container with id 2a715aa0e7459de84446d58d08828b54953cb8242a142c7dc7d71b772f2879e8 Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.033049 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hq8kp"] Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.051765 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.068723 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.299738 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.320055 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-4htn5"] Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.453014 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-nb\") pod \"7428b8ac-84e1-4438-944f-e98f085f8055\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.453267 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb4wd\" (UniqueName: \"kubernetes.io/projected/7428b8ac-84e1-4438-944f-e98f085f8055-kube-api-access-sb4wd\") pod \"7428b8ac-84e1-4438-944f-e98f085f8055\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.453416 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-dns-svc\") pod \"7428b8ac-84e1-4438-944f-e98f085f8055\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.453487 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-config\") pod \"7428b8ac-84e1-4438-944f-e98f085f8055\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.453562 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-sb\") pod \"7428b8ac-84e1-4438-944f-e98f085f8055\" (UID: \"7428b8ac-84e1-4438-944f-e98f085f8055\") " Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.476615 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7428b8ac-84e1-4438-944f-e98f085f8055-kube-api-access-sb4wd" (OuterVolumeSpecName: "kube-api-access-sb4wd") pod "7428b8ac-84e1-4438-944f-e98f085f8055" (UID: "7428b8ac-84e1-4438-944f-e98f085f8055"). InnerVolumeSpecName "kube-api-access-sb4wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.479505 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7428b8ac-84e1-4438-944f-e98f085f8055" (UID: "7428b8ac-84e1-4438-944f-e98f085f8055"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.484900 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-config" (OuterVolumeSpecName: "config") pod "7428b8ac-84e1-4438-944f-e98f085f8055" (UID: "7428b8ac-84e1-4438-944f-e98f085f8055"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.505109 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7428b8ac-84e1-4438-944f-e98f085f8055" (UID: "7428b8ac-84e1-4438-944f-e98f085f8055"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.509777 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7428b8ac-84e1-4438-944f-e98f085f8055" (UID: "7428b8ac-84e1-4438-944f-e98f085f8055"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.565492 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.565892 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb4wd\" (UniqueName: \"kubernetes.io/projected/7428b8ac-84e1-4438-944f-e98f085f8055-kube-api-access-sb4wd\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.565908 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.565917 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.565924 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7428b8ac-84e1-4438-944f-e98f085f8055-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.615817 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:09 crc kubenswrapper[4669]: W1010 09:28:09.636352 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f64ef22_59f4_471b_9726_b3673ba3ddcf.slice/crio-dfbcdbfe73307042dc35f74d8eb5581c2d3d24aac001e1724caa4abe5a37e595 WatchSource:0}: Error finding container dfbcdbfe73307042dc35f74d8eb5581c2d3d24aac001e1724caa4abe5a37e595: Status 404 returned error can't find the container with id dfbcdbfe73307042dc35f74d8eb5581c2d3d24aac001e1724caa4abe5a37e595 Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.699246 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-mcjwv"] Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.746680 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z8vqt"] Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.883698 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerStarted","Data":"dfbcdbfe73307042dc35f74d8eb5581c2d3d24aac001e1724caa4abe5a37e595"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.885449 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" event={"ID":"7428b8ac-84e1-4438-944f-e98f085f8055","Type":"ContainerDied","Data":"538fc3fe9508bdfa851dd98df70a1b51225ca8ea77bca7928af92675efd19448"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.885513 4669 scope.go:117] "RemoveContainer" containerID="4125334e23c9f137fe1abec077087863f6965bcbba1ff246fe71af2515f654c5" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.885466 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bspk7" Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.887888 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" event={"ID":"60f556bd-2a52-4976-8709-0d79392c793e","Type":"ContainerStarted","Data":"6d295bbf5add68958b27cea84353a0d9dc833780f5002ed66cc82f54b341d7e9"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.889650 4669 generic.go:334] "Generic (PLEG): container finished" podID="139fea85-7c6e-45de-b9f9-d667c361a9d1" containerID="2c0db75e9f2a78c7f5a067a975725740daa901ffa48222fdab9b9025be3800b1" exitCode=0 Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.889691 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" event={"ID":"139fea85-7c6e-45de-b9f9-d667c361a9d1","Type":"ContainerDied","Data":"2c0db75e9f2a78c7f5a067a975725740daa901ffa48222fdab9b9025be3800b1"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.889717 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" event={"ID":"139fea85-7c6e-45de-b9f9-d667c361a9d1","Type":"ContainerStarted","Data":"13711542eff588b60aa03025c36cd98bd8d01f36f63a063742ac6416c9528169"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.893339 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z8vqt" event={"ID":"799cf433-723c-4168-8522-773ca2947ec3","Type":"ContainerStarted","Data":"c60bef242fcc68f3af1cd131243a4910107e699c0a11e522bb5c1123dbefff77"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.899559 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hq8kp" event={"ID":"61c92eb8-effe-434f-9c71-cb74467723e9","Type":"ContainerStarted","Data":"e5b0a3c74455aca34a9a12e3eafd45bb2cd510b1eae7a325e0ae3ff191bc58e0"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.899712 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hq8kp" event={"ID":"61c92eb8-effe-434f-9c71-cb74467723e9","Type":"ContainerStarted","Data":"2a715aa0e7459de84446d58d08828b54953cb8242a142c7dc7d71b772f2879e8"} Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.950484 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bspk7"] Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.971401 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bspk7"] Oct 10 09:28:09 crc kubenswrapper[4669]: I1010 09:28:09.983108 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hq8kp" podStartSLOduration=1.983086004 podStartE2EDuration="1.983086004s" podCreationTimestamp="2025-10-10 09:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:09.981959248 +0000 UTC m=+1032.997978000" watchObservedRunningTime="2025-10-10 09:28:09.983086004 +0000 UTC m=+1032.999104756" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.365401 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.498501 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-nb\") pod \"139fea85-7c6e-45de-b9f9-d667c361a9d1\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.498634 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-sb\") pod \"139fea85-7c6e-45de-b9f9-d667c361a9d1\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.498654 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-dns-svc\") pod \"139fea85-7c6e-45de-b9f9-d667c361a9d1\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.498791 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb2bt\" (UniqueName: \"kubernetes.io/projected/139fea85-7c6e-45de-b9f9-d667c361a9d1-kube-api-access-tb2bt\") pod \"139fea85-7c6e-45de-b9f9-d667c361a9d1\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.498870 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-config\") pod \"139fea85-7c6e-45de-b9f9-d667c361a9d1\" (UID: \"139fea85-7c6e-45de-b9f9-d667c361a9d1\") " Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.520567 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/139fea85-7c6e-45de-b9f9-d667c361a9d1-kube-api-access-tb2bt" (OuterVolumeSpecName: "kube-api-access-tb2bt") pod "139fea85-7c6e-45de-b9f9-d667c361a9d1" (UID: "139fea85-7c6e-45de-b9f9-d667c361a9d1"). InnerVolumeSpecName "kube-api-access-tb2bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.522447 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "139fea85-7c6e-45de-b9f9-d667c361a9d1" (UID: "139fea85-7c6e-45de-b9f9-d667c361a9d1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.525182 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-config" (OuterVolumeSpecName: "config") pod "139fea85-7c6e-45de-b9f9-d667c361a9d1" (UID: "139fea85-7c6e-45de-b9f9-d667c361a9d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.530855 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "139fea85-7c6e-45de-b9f9-d667c361a9d1" (UID: "139fea85-7c6e-45de-b9f9-d667c361a9d1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.534771 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "139fea85-7c6e-45de-b9f9-d667c361a9d1" (UID: "139fea85-7c6e-45de-b9f9-d667c361a9d1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.601046 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.601085 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.601095 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.601105 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb2bt\" (UniqueName: \"kubernetes.io/projected/139fea85-7c6e-45de-b9f9-d667c361a9d1-kube-api-access-tb2bt\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.601116 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139fea85-7c6e-45de-b9f9-d667c361a9d1-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.914827 4669 generic.go:334] "Generic (PLEG): container finished" podID="60f556bd-2a52-4976-8709-0d79392c793e" containerID="4561a054c992e11cf2817733da92c62f59bd37964a58898b0025e6449112fd32" exitCode=0 Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.914888 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" event={"ID":"60f556bd-2a52-4976-8709-0d79392c793e","Type":"ContainerDied","Data":"4561a054c992e11cf2817733da92c62f59bd37964a58898b0025e6449112fd32"} Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.943710 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" event={"ID":"139fea85-7c6e-45de-b9f9-d667c361a9d1","Type":"ContainerDied","Data":"13711542eff588b60aa03025c36cd98bd8d01f36f63a063742ac6416c9528169"} Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.943761 4669 scope.go:117] "RemoveContainer" containerID="2c0db75e9f2a78c7f5a067a975725740daa901ffa48222fdab9b9025be3800b1" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.943860 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-4htn5" Oct 10 09:28:10 crc kubenswrapper[4669]: I1010 09:28:10.947484 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:11 crc kubenswrapper[4669]: I1010 09:28:11.040807 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-4htn5"] Oct 10 09:28:11 crc kubenswrapper[4669]: I1010 09:28:11.069533 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-4htn5"] Oct 10 09:28:11 crc kubenswrapper[4669]: I1010 09:28:11.820420 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="139fea85-7c6e-45de-b9f9-d667c361a9d1" path="/var/lib/kubelet/pods/139fea85-7c6e-45de-b9f9-d667c361a9d1/volumes" Oct 10 09:28:11 crc kubenswrapper[4669]: I1010 09:28:11.821266 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7428b8ac-84e1-4438-944f-e98f085f8055" path="/var/lib/kubelet/pods/7428b8ac-84e1-4438-944f-e98f085f8055/volumes" Oct 10 09:28:11 crc kubenswrapper[4669]: I1010 09:28:11.967236 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" event={"ID":"60f556bd-2a52-4976-8709-0d79392c793e","Type":"ContainerStarted","Data":"054286cb2a84d97a0999fb90f43bed134302206b4956068b9a4b7e2292dd530a"} Oct 10 09:28:11 crc kubenswrapper[4669]: I1010 09:28:11.968677 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:12 crc kubenswrapper[4669]: I1010 09:28:12.038815 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" podStartSLOduration=4.038787185 podStartE2EDuration="4.038787185s" podCreationTimestamp="2025-10-10 09:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:12.013793168 +0000 UTC m=+1035.029811910" watchObservedRunningTime="2025-10-10 09:28:12.038787185 +0000 UTC m=+1035.054805927" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.459756 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5a0a-account-create-pbfzv"] Oct 10 09:28:13 crc kubenswrapper[4669]: E1010 09:28:13.460518 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139fea85-7c6e-45de-b9f9-d667c361a9d1" containerName="init" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.460529 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="139fea85-7c6e-45de-b9f9-d667c361a9d1" containerName="init" Oct 10 09:28:13 crc kubenswrapper[4669]: E1010 09:28:13.460555 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7428b8ac-84e1-4438-944f-e98f085f8055" containerName="init" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.460562 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7428b8ac-84e1-4438-944f-e98f085f8055" containerName="init" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.460732 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7428b8ac-84e1-4438-944f-e98f085f8055" containerName="init" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.460759 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="139fea85-7c6e-45de-b9f9-d667c361a9d1" containerName="init" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.461329 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a0a-account-create-pbfzv" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.463264 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.485532 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a0a-account-create-pbfzv"] Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.550681 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-af9a-account-create-hhvnc"] Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.551914 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af9a-account-create-hhvnc" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.554958 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.555696 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kg8m\" (UniqueName: \"kubernetes.io/projected/00ebe613-c446-4805-800e-a2ce24c40bf5-kube-api-access-7kg8m\") pod \"barbican-5a0a-account-create-pbfzv\" (UID: \"00ebe613-c446-4805-800e-a2ce24c40bf5\") " pod="openstack/barbican-5a0a-account-create-pbfzv" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.564746 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-af9a-account-create-hhvnc"] Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.657236 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp5b6\" (UniqueName: \"kubernetes.io/projected/d4512bc9-9144-4bbe-8ab4-7bdb096f18aa-kube-api-access-bp5b6\") pod \"cinder-af9a-account-create-hhvnc\" (UID: \"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa\") " pod="openstack/cinder-af9a-account-create-hhvnc" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.657283 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kg8m\" (UniqueName: \"kubernetes.io/projected/00ebe613-c446-4805-800e-a2ce24c40bf5-kube-api-access-7kg8m\") pod \"barbican-5a0a-account-create-pbfzv\" (UID: \"00ebe613-c446-4805-800e-a2ce24c40bf5\") " pod="openstack/barbican-5a0a-account-create-pbfzv" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.685970 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kg8m\" (UniqueName: \"kubernetes.io/projected/00ebe613-c446-4805-800e-a2ce24c40bf5-kube-api-access-7kg8m\") pod \"barbican-5a0a-account-create-pbfzv\" (UID: \"00ebe613-c446-4805-800e-a2ce24c40bf5\") " pod="openstack/barbican-5a0a-account-create-pbfzv" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.759005 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp5b6\" (UniqueName: \"kubernetes.io/projected/d4512bc9-9144-4bbe-8ab4-7bdb096f18aa-kube-api-access-bp5b6\") pod \"cinder-af9a-account-create-hhvnc\" (UID: \"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa\") " pod="openstack/cinder-af9a-account-create-hhvnc" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.778112 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a0a-account-create-pbfzv" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.791974 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp5b6\" (UniqueName: \"kubernetes.io/projected/d4512bc9-9144-4bbe-8ab4-7bdb096f18aa-kube-api-access-bp5b6\") pod \"cinder-af9a-account-create-hhvnc\" (UID: \"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa\") " pod="openstack/cinder-af9a-account-create-hhvnc" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.847853 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-4851-account-create-lctwn"] Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.848939 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4851-account-create-lctwn" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.852036 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.866838 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-4851-account-create-lctwn"] Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.905189 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af9a-account-create-hhvnc" Oct 10 09:28:13 crc kubenswrapper[4669]: I1010 09:28:13.963029 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9gb8\" (UniqueName: \"kubernetes.io/projected/bf031e52-cb5b-49be-be4f-c136299774fe-kube-api-access-x9gb8\") pod \"neutron-4851-account-create-lctwn\" (UID: \"bf031e52-cb5b-49be-be4f-c136299774fe\") " pod="openstack/neutron-4851-account-create-lctwn" Oct 10 09:28:14 crc kubenswrapper[4669]: I1010 09:28:14.003737 4669 generic.go:334] "Generic (PLEG): container finished" podID="61c92eb8-effe-434f-9c71-cb74467723e9" containerID="e5b0a3c74455aca34a9a12e3eafd45bb2cd510b1eae7a325e0ae3ff191bc58e0" exitCode=0 Oct 10 09:28:14 crc kubenswrapper[4669]: I1010 09:28:14.003795 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hq8kp" event={"ID":"61c92eb8-effe-434f-9c71-cb74467723e9","Type":"ContainerDied","Data":"e5b0a3c74455aca34a9a12e3eafd45bb2cd510b1eae7a325e0ae3ff191bc58e0"} Oct 10 09:28:14 crc kubenswrapper[4669]: I1010 09:28:14.064567 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9gb8\" (UniqueName: \"kubernetes.io/projected/bf031e52-cb5b-49be-be4f-c136299774fe-kube-api-access-x9gb8\") pod \"neutron-4851-account-create-lctwn\" (UID: \"bf031e52-cb5b-49be-be4f-c136299774fe\") " pod="openstack/neutron-4851-account-create-lctwn" Oct 10 09:28:14 crc kubenswrapper[4669]: I1010 09:28:14.086453 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9gb8\" (UniqueName: \"kubernetes.io/projected/bf031e52-cb5b-49be-be4f-c136299774fe-kube-api-access-x9gb8\") pod \"neutron-4851-account-create-lctwn\" (UID: \"bf031e52-cb5b-49be-be4f-c136299774fe\") " pod="openstack/neutron-4851-account-create-lctwn" Oct 10 09:28:14 crc kubenswrapper[4669]: I1010 09:28:14.181759 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4851-account-create-lctwn" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.775048 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.840987 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-combined-ca-bundle\") pod \"61c92eb8-effe-434f-9c71-cb74467723e9\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.841077 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t4st\" (UniqueName: \"kubernetes.io/projected/61c92eb8-effe-434f-9c71-cb74467723e9-kube-api-access-7t4st\") pod \"61c92eb8-effe-434f-9c71-cb74467723e9\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.841167 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-fernet-keys\") pod \"61c92eb8-effe-434f-9c71-cb74467723e9\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.841224 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-credential-keys\") pod \"61c92eb8-effe-434f-9c71-cb74467723e9\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.841244 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-config-data\") pod \"61c92eb8-effe-434f-9c71-cb74467723e9\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.841275 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-scripts\") pod \"61c92eb8-effe-434f-9c71-cb74467723e9\" (UID: \"61c92eb8-effe-434f-9c71-cb74467723e9\") " Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.849994 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "61c92eb8-effe-434f-9c71-cb74467723e9" (UID: "61c92eb8-effe-434f-9c71-cb74467723e9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.851755 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c92eb8-effe-434f-9c71-cb74467723e9-kube-api-access-7t4st" (OuterVolumeSpecName: "kube-api-access-7t4st") pod "61c92eb8-effe-434f-9c71-cb74467723e9" (UID: "61c92eb8-effe-434f-9c71-cb74467723e9"). InnerVolumeSpecName "kube-api-access-7t4st". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.852955 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "61c92eb8-effe-434f-9c71-cb74467723e9" (UID: "61c92eb8-effe-434f-9c71-cb74467723e9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.854098 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-scripts" (OuterVolumeSpecName: "scripts") pod "61c92eb8-effe-434f-9c71-cb74467723e9" (UID: "61c92eb8-effe-434f-9c71-cb74467723e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.895781 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61c92eb8-effe-434f-9c71-cb74467723e9" (UID: "61c92eb8-effe-434f-9c71-cb74467723e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.895841 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-config-data" (OuterVolumeSpecName: "config-data") pod "61c92eb8-effe-434f-9c71-cb74467723e9" (UID: "61c92eb8-effe-434f-9c71-cb74467723e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.943040 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.943069 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.943080 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t4st\" (UniqueName: \"kubernetes.io/projected/61c92eb8-effe-434f-9c71-cb74467723e9-kube-api-access-7t4st\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.943088 4669 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.943096 4669 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:16 crc kubenswrapper[4669]: I1010 09:28:16.943104 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61c92eb8-effe-434f-9c71-cb74467723e9-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.053689 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z8vqt" event={"ID":"799cf433-723c-4168-8522-773ca2947ec3","Type":"ContainerStarted","Data":"d778f8e40a75b782c3fcb763fc57410d9ffc558284a07569abfe8a957d75a712"} Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.055776 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hq8kp" event={"ID":"61c92eb8-effe-434f-9c71-cb74467723e9","Type":"ContainerDied","Data":"2a715aa0e7459de84446d58d08828b54953cb8242a142c7dc7d71b772f2879e8"} Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.055798 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a715aa0e7459de84446d58d08828b54953cb8242a142c7dc7d71b772f2879e8" Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.055833 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hq8kp" Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.066960 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerStarted","Data":"32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49"} Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.112763 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-z8vqt" podStartSLOduration=2.246711317 podStartE2EDuration="9.112746123s" podCreationTimestamp="2025-10-10 09:28:08 +0000 UTC" firstStartedPulling="2025-10-10 09:28:09.765340018 +0000 UTC m=+1032.781358750" lastFinishedPulling="2025-10-10 09:28:16.631374814 +0000 UTC m=+1039.647393556" observedRunningTime="2025-10-10 09:28:17.084505725 +0000 UTC m=+1040.100524467" watchObservedRunningTime="2025-10-10 09:28:17.112746123 +0000 UTC m=+1040.128764865" Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.113521 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-af9a-account-create-hhvnc"] Oct 10 09:28:17 crc kubenswrapper[4669]: W1010 09:28:17.114440 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4512bc9_9144_4bbe_8ab4_7bdb096f18aa.slice/crio-3beed0a412d112afb1e4deedfc7c4bec582826b7e7313635ed9e9ed4017b8da8 WatchSource:0}: Error finding container 3beed0a412d112afb1e4deedfc7c4bec582826b7e7313635ed9e9ed4017b8da8: Status 404 returned error can't find the container with id 3beed0a412d112afb1e4deedfc7c4bec582826b7e7313635ed9e9ed4017b8da8 Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.200522 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5a0a-account-create-pbfzv"] Oct 10 09:28:17 crc kubenswrapper[4669]: W1010 09:28:17.201149 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00ebe613_c446_4805_800e_a2ce24c40bf5.slice/crio-88189346b0cded1f66ab481b587233c16f3661ccdff81d51b991e7aa40a4764a WatchSource:0}: Error finding container 88189346b0cded1f66ab481b587233c16f3661ccdff81d51b991e7aa40a4764a: Status 404 returned error can't find the container with id 88189346b0cded1f66ab481b587233c16f3661ccdff81d51b991e7aa40a4764a Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.225816 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-4851-account-create-lctwn"] Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.928730 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hq8kp"] Oct 10 09:28:17 crc kubenswrapper[4669]: I1010 09:28:17.940572 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hq8kp"] Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.042934 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-lm4k9"] Oct 10 09:28:18 crc kubenswrapper[4669]: E1010 09:28:18.044000 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c92eb8-effe-434f-9c71-cb74467723e9" containerName="keystone-bootstrap" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.044183 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c92eb8-effe-434f-9c71-cb74467723e9" containerName="keystone-bootstrap" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.054902 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c92eb8-effe-434f-9c71-cb74467723e9" containerName="keystone-bootstrap" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.056250 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.061131 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.061207 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-r86r4" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.061405 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.064805 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.100229 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lm4k9"] Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.138190 4669 generic.go:334] "Generic (PLEG): container finished" podID="00ebe613-c446-4805-800e-a2ce24c40bf5" containerID="10124d233567fa84e3a8db93abb7b4b87d2aee9f10df8b204cb7a79c0213aec0" exitCode=0 Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.141162 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a0a-account-create-pbfzv" event={"ID":"00ebe613-c446-4805-800e-a2ce24c40bf5","Type":"ContainerDied","Data":"10124d233567fa84e3a8db93abb7b4b87d2aee9f10df8b204cb7a79c0213aec0"} Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.141204 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a0a-account-create-pbfzv" event={"ID":"00ebe613-c446-4805-800e-a2ce24c40bf5","Type":"ContainerStarted","Data":"88189346b0cded1f66ab481b587233c16f3661ccdff81d51b991e7aa40a4764a"} Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.143600 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf031e52-cb5b-49be-be4f-c136299774fe" containerID="5897be1a6b0542d0b527f840ebc2861fa37c672cec4e9488e2ae8dc51bfcb40a" exitCode=0 Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.143659 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4851-account-create-lctwn" event={"ID":"bf031e52-cb5b-49be-be4f-c136299774fe","Type":"ContainerDied","Data":"5897be1a6b0542d0b527f840ebc2861fa37c672cec4e9488e2ae8dc51bfcb40a"} Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.143683 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4851-account-create-lctwn" event={"ID":"bf031e52-cb5b-49be-be4f-c136299774fe","Type":"ContainerStarted","Data":"246bdd088a67a7c7b886db10b1932a683bcebc4b70660d96080ae813f50be329"} Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.145779 4669 generic.go:334] "Generic (PLEG): container finished" podID="d4512bc9-9144-4bbe-8ab4-7bdb096f18aa" containerID="19a34cf929c657db2e9a3c7f84851bfeae4662057f442aee63d4136bcc10fb15" exitCode=0 Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.146455 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-af9a-account-create-hhvnc" event={"ID":"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa","Type":"ContainerDied","Data":"19a34cf929c657db2e9a3c7f84851bfeae4662057f442aee63d4136bcc10fb15"} Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.146482 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-af9a-account-create-hhvnc" event={"ID":"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa","Type":"ContainerStarted","Data":"3beed0a412d112afb1e4deedfc7c4bec582826b7e7313635ed9e9ed4017b8da8"} Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.175958 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-scripts\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.176002 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-fernet-keys\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.176042 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-credential-keys\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.176075 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrpvx\" (UniqueName: \"kubernetes.io/projected/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-kube-api-access-mrpvx\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.176100 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-combined-ca-bundle\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.176119 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-config-data\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.278766 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-combined-ca-bundle\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.278822 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-config-data\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.278903 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-scripts\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.278929 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-fernet-keys\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.278970 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-credential-keys\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.279006 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrpvx\" (UniqueName: \"kubernetes.io/projected/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-kube-api-access-mrpvx\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.285884 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-scripts\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.291841 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-combined-ca-bundle\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.291928 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-credential-keys\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.292377 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-fernet-keys\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.293414 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-config-data\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.299647 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrpvx\" (UniqueName: \"kubernetes.io/projected/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-kube-api-access-mrpvx\") pod \"keystone-bootstrap-lm4k9\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.404196 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:18 crc kubenswrapper[4669]: I1010 09:28:18.951497 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lm4k9"] Oct 10 09:28:18 crc kubenswrapper[4669]: W1010 09:28:18.957324 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ea3c637_dcb5_4b7c_827f_f673f9d7d8d1.slice/crio-a57ab5bc49b6b243fb106212bced870e961beb41ace39b7479f304c348e9d172 WatchSource:0}: Error finding container a57ab5bc49b6b243fb106212bced870e961beb41ace39b7479f304c348e9d172: Status 404 returned error can't find the container with id a57ab5bc49b6b243fb106212bced870e961beb41ace39b7479f304c348e9d172 Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.053958 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.135564 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-9bvgm"] Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.135808 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" podUID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerName="dnsmasq-dns" containerID="cri-o://91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b" gracePeriod=10 Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.153974 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerStarted","Data":"f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810"} Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.154887 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lm4k9" event={"ID":"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1","Type":"ContainerStarted","Data":"a57ab5bc49b6b243fb106212bced870e961beb41ace39b7479f304c348e9d172"} Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.810179 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a0a-account-create-pbfzv" Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.814657 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c92eb8-effe-434f-9c71-cb74467723e9" path="/var/lib/kubelet/pods/61c92eb8-effe-434f-9c71-cb74467723e9/volumes" Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.915979 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af9a-account-create-hhvnc" Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.923802 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4851-account-create-lctwn" Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.935699 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kg8m\" (UniqueName: \"kubernetes.io/projected/00ebe613-c446-4805-800e-a2ce24c40bf5-kube-api-access-7kg8m\") pod \"00ebe613-c446-4805-800e-a2ce24c40bf5\" (UID: \"00ebe613-c446-4805-800e-a2ce24c40bf5\") " Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.942279 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ebe613-c446-4805-800e-a2ce24c40bf5-kube-api-access-7kg8m" (OuterVolumeSpecName: "kube-api-access-7kg8m") pod "00ebe613-c446-4805-800e-a2ce24c40bf5" (UID: "00ebe613-c446-4805-800e-a2ce24c40bf5"). InnerVolumeSpecName "kube-api-access-7kg8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:19 crc kubenswrapper[4669]: I1010 09:28:19.946424 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.037528 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-nb\") pod \"a5bce784-8452-4b77-a1d0-539e471aed1c\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.037755 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp5b6\" (UniqueName: \"kubernetes.io/projected/d4512bc9-9144-4bbe-8ab4-7bdb096f18aa-kube-api-access-bp5b6\") pod \"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa\" (UID: \"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa\") " Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.037805 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9gb8\" (UniqueName: \"kubernetes.io/projected/bf031e52-cb5b-49be-be4f-c136299774fe-kube-api-access-x9gb8\") pod \"bf031e52-cb5b-49be-be4f-c136299774fe\" (UID: \"bf031e52-cb5b-49be-be4f-c136299774fe\") " Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.037836 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khpkp\" (UniqueName: \"kubernetes.io/projected/a5bce784-8452-4b77-a1d0-539e471aed1c-kube-api-access-khpkp\") pod \"a5bce784-8452-4b77-a1d0-539e471aed1c\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.037901 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-sb\") pod \"a5bce784-8452-4b77-a1d0-539e471aed1c\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.037932 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-config\") pod \"a5bce784-8452-4b77-a1d0-539e471aed1c\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.037958 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-dns-svc\") pod \"a5bce784-8452-4b77-a1d0-539e471aed1c\" (UID: \"a5bce784-8452-4b77-a1d0-539e471aed1c\") " Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.038257 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kg8m\" (UniqueName: \"kubernetes.io/projected/00ebe613-c446-4805-800e-a2ce24c40bf5-kube-api-access-7kg8m\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.048903 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf031e52-cb5b-49be-be4f-c136299774fe-kube-api-access-x9gb8" (OuterVolumeSpecName: "kube-api-access-x9gb8") pod "bf031e52-cb5b-49be-be4f-c136299774fe" (UID: "bf031e52-cb5b-49be-be4f-c136299774fe"). InnerVolumeSpecName "kube-api-access-x9gb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.048954 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5bce784-8452-4b77-a1d0-539e471aed1c-kube-api-access-khpkp" (OuterVolumeSpecName: "kube-api-access-khpkp") pod "a5bce784-8452-4b77-a1d0-539e471aed1c" (UID: "a5bce784-8452-4b77-a1d0-539e471aed1c"). InnerVolumeSpecName "kube-api-access-khpkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.051720 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4512bc9-9144-4bbe-8ab4-7bdb096f18aa-kube-api-access-bp5b6" (OuterVolumeSpecName: "kube-api-access-bp5b6") pod "d4512bc9-9144-4bbe-8ab4-7bdb096f18aa" (UID: "d4512bc9-9144-4bbe-8ab4-7bdb096f18aa"). InnerVolumeSpecName "kube-api-access-bp5b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.095261 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a5bce784-8452-4b77-a1d0-539e471aed1c" (UID: "a5bce784-8452-4b77-a1d0-539e471aed1c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.095904 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a5bce784-8452-4b77-a1d0-539e471aed1c" (UID: "a5bce784-8452-4b77-a1d0-539e471aed1c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.107662 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a5bce784-8452-4b77-a1d0-539e471aed1c" (UID: "a5bce784-8452-4b77-a1d0-539e471aed1c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.115261 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-config" (OuterVolumeSpecName: "config") pod "a5bce784-8452-4b77-a1d0-539e471aed1c" (UID: "a5bce784-8452-4b77-a1d0-539e471aed1c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.139953 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.139992 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp5b6\" (UniqueName: \"kubernetes.io/projected/d4512bc9-9144-4bbe-8ab4-7bdb096f18aa-kube-api-access-bp5b6\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.140010 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9gb8\" (UniqueName: \"kubernetes.io/projected/bf031e52-cb5b-49be-be4f-c136299774fe-kube-api-access-x9gb8\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.140023 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khpkp\" (UniqueName: \"kubernetes.io/projected/a5bce784-8452-4b77-a1d0-539e471aed1c-kube-api-access-khpkp\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.140033 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.140043 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.140051 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a5bce784-8452-4b77-a1d0-539e471aed1c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.165144 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lm4k9" event={"ID":"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1","Type":"ContainerStarted","Data":"ee1036c5e473647e3b0c63ba754ff95245c3d11d88512f8e661b954db1f591c0"} Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.169695 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5a0a-account-create-pbfzv" event={"ID":"00ebe613-c446-4805-800e-a2ce24c40bf5","Type":"ContainerDied","Data":"88189346b0cded1f66ab481b587233c16f3661ccdff81d51b991e7aa40a4764a"} Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.169731 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88189346b0cded1f66ab481b587233c16f3661ccdff81d51b991e7aa40a4764a" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.169781 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5a0a-account-create-pbfzv" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.188709 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-lm4k9" podStartSLOduration=2.188694421 podStartE2EDuration="2.188694421s" podCreationTimestamp="2025-10-10 09:28:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:20.180770707 +0000 UTC m=+1043.196789449" watchObservedRunningTime="2025-10-10 09:28:20.188694421 +0000 UTC m=+1043.204713163" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.190065 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4851-account-create-lctwn" event={"ID":"bf031e52-cb5b-49be-be4f-c136299774fe","Type":"ContainerDied","Data":"246bdd088a67a7c7b886db10b1932a683bcebc4b70660d96080ae813f50be329"} Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.190096 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="246bdd088a67a7c7b886db10b1932a683bcebc4b70660d96080ae813f50be329" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.190157 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4851-account-create-lctwn" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.198833 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-af9a-account-create-hhvnc" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.198856 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-af9a-account-create-hhvnc" event={"ID":"d4512bc9-9144-4bbe-8ab4-7bdb096f18aa","Type":"ContainerDied","Data":"3beed0a412d112afb1e4deedfc7c4bec582826b7e7313635ed9e9ed4017b8da8"} Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.198892 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3beed0a412d112afb1e4deedfc7c4bec582826b7e7313635ed9e9ed4017b8da8" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.201079 4669 generic.go:334] "Generic (PLEG): container finished" podID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerID="91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b" exitCode=0 Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.201132 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" event={"ID":"a5bce784-8452-4b77-a1d0-539e471aed1c","Type":"ContainerDied","Data":"91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b"} Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.201152 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" event={"ID":"a5bce784-8452-4b77-a1d0-539e471aed1c","Type":"ContainerDied","Data":"981ba3aebe4610cf6d68bc6bd0fd07d06350a0ad7d989579e2a95a460e807765"} Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.201170 4669 scope.go:117] "RemoveContainer" containerID="91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.201278 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-9bvgm" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.216217 4669 generic.go:334] "Generic (PLEG): container finished" podID="799cf433-723c-4168-8522-773ca2947ec3" containerID="d778f8e40a75b782c3fcb763fc57410d9ffc558284a07569abfe8a957d75a712" exitCode=0 Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.216257 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z8vqt" event={"ID":"799cf433-723c-4168-8522-773ca2947ec3","Type":"ContainerDied","Data":"d778f8e40a75b782c3fcb763fc57410d9ffc558284a07569abfe8a957d75a712"} Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.255073 4669 scope.go:117] "RemoveContainer" containerID="5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.259476 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-9bvgm"] Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.265738 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-9bvgm"] Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.294261 4669 scope.go:117] "RemoveContainer" containerID="91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b" Oct 10 09:28:20 crc kubenswrapper[4669]: E1010 09:28:20.294838 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b\": container with ID starting with 91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b not found: ID does not exist" containerID="91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.294865 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b"} err="failed to get container status \"91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b\": rpc error: code = NotFound desc = could not find container \"91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b\": container with ID starting with 91cf57198dc7c9ce35d6e91079c92910cc82678d8674c9b2f3b1fc39e1bd023b not found: ID does not exist" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.294884 4669 scope.go:117] "RemoveContainer" containerID="5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e" Oct 10 09:28:20 crc kubenswrapper[4669]: E1010 09:28:20.295404 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e\": container with ID starting with 5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e not found: ID does not exist" containerID="5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e" Oct 10 09:28:20 crc kubenswrapper[4669]: I1010 09:28:20.295456 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e"} err="failed to get container status \"5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e\": rpc error: code = NotFound desc = could not find container \"5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e\": container with ID starting with 5adb0a905c1c0a5cbf87704c75d8f5471bb0b659749925a51d1b3f69f935020e not found: ID does not exist" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.535378 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.689347 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-combined-ca-bundle\") pod \"799cf433-723c-4168-8522-773ca2947ec3\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.689520 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glkr2\" (UniqueName: \"kubernetes.io/projected/799cf433-723c-4168-8522-773ca2947ec3-kube-api-access-glkr2\") pod \"799cf433-723c-4168-8522-773ca2947ec3\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.689631 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-scripts\") pod \"799cf433-723c-4168-8522-773ca2947ec3\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.689671 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/799cf433-723c-4168-8522-773ca2947ec3-logs\") pod \"799cf433-723c-4168-8522-773ca2947ec3\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.689763 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-config-data\") pod \"799cf433-723c-4168-8522-773ca2947ec3\" (UID: \"799cf433-723c-4168-8522-773ca2947ec3\") " Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.691097 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/799cf433-723c-4168-8522-773ca2947ec3-logs" (OuterVolumeSpecName: "logs") pod "799cf433-723c-4168-8522-773ca2947ec3" (UID: "799cf433-723c-4168-8522-773ca2947ec3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.696068 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-scripts" (OuterVolumeSpecName: "scripts") pod "799cf433-723c-4168-8522-773ca2947ec3" (UID: "799cf433-723c-4168-8522-773ca2947ec3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.709115 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/799cf433-723c-4168-8522-773ca2947ec3-kube-api-access-glkr2" (OuterVolumeSpecName: "kube-api-access-glkr2") pod "799cf433-723c-4168-8522-773ca2947ec3" (UID: "799cf433-723c-4168-8522-773ca2947ec3"). InnerVolumeSpecName "kube-api-access-glkr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.726085 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-config-data" (OuterVolumeSpecName: "config-data") pod "799cf433-723c-4168-8522-773ca2947ec3" (UID: "799cf433-723c-4168-8522-773ca2947ec3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.729139 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "799cf433-723c-4168-8522-773ca2947ec3" (UID: "799cf433-723c-4168-8522-773ca2947ec3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.791390 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.791422 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/799cf433-723c-4168-8522-773ca2947ec3-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.791432 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.791441 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/799cf433-723c-4168-8522-773ca2947ec3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.791453 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glkr2\" (UniqueName: \"kubernetes.io/projected/799cf433-723c-4168-8522-773ca2947ec3-kube-api-access-glkr2\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:21 crc kubenswrapper[4669]: I1010 09:28:21.805133 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5bce784-8452-4b77-a1d0-539e471aed1c" path="/var/lib/kubelet/pods/a5bce784-8452-4b77-a1d0-539e471aed1c/volumes" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.272794 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z8vqt" event={"ID":"799cf433-723c-4168-8522-773ca2947ec3","Type":"ContainerDied","Data":"c60bef242fcc68f3af1cd131243a4910107e699c0a11e522bb5c1123dbefff77"} Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.272832 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c60bef242fcc68f3af1cd131243a4910107e699c0a11e522bb5c1123dbefff77" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.272887 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z8vqt" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344175 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7f557bc54-vbfbn"] Oct 10 09:28:22 crc kubenswrapper[4669]: E1010 09:28:22.344614 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerName="init" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344635 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerName="init" Oct 10 09:28:22 crc kubenswrapper[4669]: E1010 09:28:22.344651 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="799cf433-723c-4168-8522-773ca2947ec3" containerName="placement-db-sync" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344658 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="799cf433-723c-4168-8522-773ca2947ec3" containerName="placement-db-sync" Oct 10 09:28:22 crc kubenswrapper[4669]: E1010 09:28:22.344666 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf031e52-cb5b-49be-be4f-c136299774fe" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344673 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf031e52-cb5b-49be-be4f-c136299774fe" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: E1010 09:28:22.344683 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ebe613-c446-4805-800e-a2ce24c40bf5" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344689 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ebe613-c446-4805-800e-a2ce24c40bf5" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: E1010 09:28:22.344700 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerName="dnsmasq-dns" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344705 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerName="dnsmasq-dns" Oct 10 09:28:22 crc kubenswrapper[4669]: E1010 09:28:22.344714 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4512bc9-9144-4bbe-8ab4-7bdb096f18aa" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344719 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4512bc9-9144-4bbe-8ab4-7bdb096f18aa" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344887 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4512bc9-9144-4bbe-8ab4-7bdb096f18aa" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344915 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf031e52-cb5b-49be-be4f-c136299774fe" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344936 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="799cf433-723c-4168-8522-773ca2947ec3" containerName="placement-db-sync" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344945 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ebe613-c446-4805-800e-a2ce24c40bf5" containerName="mariadb-account-create" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.344960 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5bce784-8452-4b77-a1d0-539e471aed1c" containerName="dnsmasq-dns" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.345902 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.348943 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.349091 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.349242 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-8hzxw" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.349358 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.349449 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.350812 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7f557bc54-vbfbn"] Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.502378 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9be36cc-2136-4285-8489-93149d0136c3-logs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.502748 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpp5g\" (UniqueName: \"kubernetes.io/projected/d9be36cc-2136-4285-8489-93149d0136c3-kube-api-access-xpp5g\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.502810 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-public-tls-certs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.502873 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-config-data\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.502901 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-scripts\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.503156 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-combined-ca-bundle\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.503284 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-internal-tls-certs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.605001 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-public-tls-certs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.605055 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-config-data\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.605087 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-scripts\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.605128 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-combined-ca-bundle\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.605161 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-internal-tls-certs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.605188 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9be36cc-2136-4285-8489-93149d0136c3-logs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.605205 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpp5g\" (UniqueName: \"kubernetes.io/projected/d9be36cc-2136-4285-8489-93149d0136c3-kube-api-access-xpp5g\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.606190 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9be36cc-2136-4285-8489-93149d0136c3-logs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.609706 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-config-data\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.614817 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-combined-ca-bundle\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.615166 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-public-tls-certs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.618753 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-internal-tls-certs\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.618978 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9be36cc-2136-4285-8489-93149d0136c3-scripts\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.621513 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpp5g\" (UniqueName: \"kubernetes.io/projected/d9be36cc-2136-4285-8489-93149d0136c3-kube-api-access-xpp5g\") pod \"placement-7f557bc54-vbfbn\" (UID: \"d9be36cc-2136-4285-8489-93149d0136c3\") " pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:22 crc kubenswrapper[4669]: I1010 09:28:22.688413 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:23 crc kubenswrapper[4669]: I1010 09:28:23.989174 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-trw89"] Oct 10 09:28:23 crc kubenswrapper[4669]: I1010 09:28:23.990132 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:23 crc kubenswrapper[4669]: I1010 09:28:23.993026 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 09:28:23 crc kubenswrapper[4669]: I1010 09:28:23.993262 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jpczk" Oct 10 09:28:23 crc kubenswrapper[4669]: I1010 09:28:23.993388 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.008274 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-trw89"] Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.028439 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-etc-machine-id\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.028486 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-config-data\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.028542 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t29ld\" (UniqueName: \"kubernetes.io/projected/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-kube-api-access-t29ld\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.028568 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-db-sync-config-data\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.028602 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-scripts\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.028622 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-combined-ca-bundle\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.118367 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-l7kr8"] Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.119760 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.123363 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-j5xr7" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.123504 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.128094 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-l7kr8"] Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.141321 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t29ld\" (UniqueName: \"kubernetes.io/projected/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-kube-api-access-t29ld\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.141379 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-db-sync-config-data\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.141401 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-db-sync-config-data\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.141426 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-scripts\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.141444 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-combined-ca-bundle\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.141483 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-combined-ca-bundle\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.141507 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-etc-machine-id\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.142466 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-config-data\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.142536 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-etc-machine-id\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.142493 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkrc6\" (UniqueName: \"kubernetes.io/projected/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-kube-api-access-tkrc6\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.152113 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-scripts\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.158518 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-combined-ca-bundle\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.169082 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-db-sync-config-data\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.180936 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-config-data\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.185208 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t29ld\" (UniqueName: \"kubernetes.io/projected/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-kube-api-access-t29ld\") pod \"cinder-db-sync-trw89\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.198630 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-d8m9m"] Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.199763 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.202024 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.202172 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.202287 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9zprf" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.207930 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d8m9m"] Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.243939 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkrc6\" (UniqueName: \"kubernetes.io/projected/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-kube-api-access-tkrc6\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.244017 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-db-sync-config-data\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.244071 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-combined-ca-bundle\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.254031 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-db-sync-config-data\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.262002 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkrc6\" (UniqueName: \"kubernetes.io/projected/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-kube-api-access-tkrc6\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.264418 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-combined-ca-bundle\") pod \"barbican-db-sync-l7kr8\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.291172 4669 generic.go:334] "Generic (PLEG): container finished" podID="6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" containerID="ee1036c5e473647e3b0c63ba754ff95245c3d11d88512f8e661b954db1f591c0" exitCode=0 Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.291212 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lm4k9" event={"ID":"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1","Type":"ContainerDied","Data":"ee1036c5e473647e3b0c63ba754ff95245c3d11d88512f8e661b954db1f591c0"} Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.314728 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-trw89" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.346554 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-config\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.346684 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86smk\" (UniqueName: \"kubernetes.io/projected/50102622-b6c0-4fe9-8eab-b4027da1a36a-kube-api-access-86smk\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.346770 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-combined-ca-bundle\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.448285 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86smk\" (UniqueName: \"kubernetes.io/projected/50102622-b6c0-4fe9-8eab-b4027da1a36a-kube-api-access-86smk\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.448387 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-combined-ca-bundle\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.448430 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-config\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.452300 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-combined-ca-bundle\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.470407 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-config\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.475897 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86smk\" (UniqueName: \"kubernetes.io/projected/50102622-b6c0-4fe9-8eab-b4027da1a36a-kube-api-access-86smk\") pod \"neutron-db-sync-d8m9m\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.523914 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:24 crc kubenswrapper[4669]: I1010 09:28:24.542226 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.231643 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.285556 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-scripts\") pod \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.285651 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrpvx\" (UniqueName: \"kubernetes.io/projected/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-kube-api-access-mrpvx\") pod \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.286563 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-credential-keys\") pod \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.286659 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-config-data\") pod \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.286772 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-combined-ca-bundle\") pod \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.286798 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-fernet-keys\") pod \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\" (UID: \"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1\") " Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.292789 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-scripts" (OuterVolumeSpecName: "scripts") pod "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" (UID: "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.292839 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" (UID: "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.297415 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-kube-api-access-mrpvx" (OuterVolumeSpecName: "kube-api-access-mrpvx") pod "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" (UID: "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1"). InnerVolumeSpecName "kube-api-access-mrpvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.305647 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" (UID: "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.312779 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lm4k9" event={"ID":"6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1","Type":"ContainerDied","Data":"a57ab5bc49b6b243fb106212bced870e961beb41ace39b7479f304c348e9d172"} Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.312812 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a57ab5bc49b6b243fb106212bced870e961beb41ace39b7479f304c348e9d172" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.312866 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lm4k9" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.335971 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-config-data" (OuterVolumeSpecName: "config-data") pod "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" (UID: "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.363955 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" (UID: "6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.388905 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.388962 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrpvx\" (UniqueName: \"kubernetes.io/projected/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-kube-api-access-mrpvx\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.388972 4669 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.388981 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.389092 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.389106 4669 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.466989 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-65b8b49f6-vm5h8"] Oct 10 09:28:26 crc kubenswrapper[4669]: E1010 09:28:26.467287 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" containerName="keystone-bootstrap" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.467297 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" containerName="keystone-bootstrap" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.467458 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" containerName="keystone-bootstrap" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.467949 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.475011 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.475157 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.498240 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-65b8b49f6-vm5h8"] Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513616 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-config-data\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513682 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-public-tls-certs\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513712 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-scripts\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513749 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-credential-keys\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513796 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-internal-tls-certs\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513821 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-combined-ca-bundle\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513841 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcl6w\" (UniqueName: \"kubernetes.io/projected/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-kube-api-access-kcl6w\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.513886 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-fernet-keys\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615435 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-fernet-keys\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615547 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-config-data\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615609 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-public-tls-certs\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615638 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-scripts\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615678 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-credential-keys\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615728 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-combined-ca-bundle\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615748 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-internal-tls-certs\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.615768 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcl6w\" (UniqueName: \"kubernetes.io/projected/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-kube-api-access-kcl6w\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.621743 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-scripts\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.621932 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-combined-ca-bundle\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.624053 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-fernet-keys\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.624410 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-credential-keys\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.631901 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-internal-tls-certs\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.632367 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-public-tls-certs\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.634410 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcl6w\" (UniqueName: \"kubernetes.io/projected/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-kube-api-access-kcl6w\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.635173 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c9b81a2-843b-4bab-b58a-5e2a1ce2e636-config-data\") pod \"keystone-65b8b49f6-vm5h8\" (UID: \"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636\") " pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.747307 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d8m9m"] Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.779302 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7f557bc54-vbfbn"] Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.788707 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-l7kr8"] Oct 10 09:28:26 crc kubenswrapper[4669]: W1010 09:28:26.797826 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37ed6a10_c32a_4f82_8fa6_63e0cd8de8ff.slice/crio-30b5cfbab5e9eefe0de80cab4bbe44acb0615a2c66884da7393b2afc60642333 WatchSource:0}: Error finding container 30b5cfbab5e9eefe0de80cab4bbe44acb0615a2c66884da7393b2afc60642333: Status 404 returned error can't find the container with id 30b5cfbab5e9eefe0de80cab4bbe44acb0615a2c66884da7393b2afc60642333 Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.805414 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:26 crc kubenswrapper[4669]: I1010 09:28:26.808780 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-trw89"] Oct 10 09:28:26 crc kubenswrapper[4669]: W1010 09:28:26.831776 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1d9870d_304c_4ab1_9091_9e7e939cc6b8.slice/crio-4c1ae6a70585cb9226fcfdac265a0e882b17cca33ee557c6a7b3baeee9fc2b56 WatchSource:0}: Error finding container 4c1ae6a70585cb9226fcfdac265a0e882b17cca33ee557c6a7b3baeee9fc2b56: Status 404 returned error can't find the container with id 4c1ae6a70585cb9226fcfdac265a0e882b17cca33ee557c6a7b3baeee9fc2b56 Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.256661 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-65b8b49f6-vm5h8"] Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.338195 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l7kr8" event={"ID":"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff","Type":"ContainerStarted","Data":"30b5cfbab5e9eefe0de80cab4bbe44acb0615a2c66884da7393b2afc60642333"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.341238 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerStarted","Data":"ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.344432 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d8m9m" event={"ID":"50102622-b6c0-4fe9-8eab-b4027da1a36a","Type":"ContainerStarted","Data":"416e34c6fab75954a71767e78290564c70e9e1feae973fab44258c9fa061165f"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.344473 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d8m9m" event={"ID":"50102622-b6c0-4fe9-8eab-b4027da1a36a","Type":"ContainerStarted","Data":"700baef58b6a24e63d6fc452bab60dc8f0730748c5d3e29b2854aea725ffce78"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.345549 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-trw89" event={"ID":"a1d9870d-304c-4ab1-9091-9e7e939cc6b8","Type":"ContainerStarted","Data":"4c1ae6a70585cb9226fcfdac265a0e882b17cca33ee557c6a7b3baeee9fc2b56"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.348234 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f557bc54-vbfbn" event={"ID":"d9be36cc-2136-4285-8489-93149d0136c3","Type":"ContainerStarted","Data":"858705f86ca35420a142ebcf6c878e75871172dbe4437f32bfeff5f1b0500c2a"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.348284 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f557bc54-vbfbn" event={"ID":"d9be36cc-2136-4285-8489-93149d0136c3","Type":"ContainerStarted","Data":"4c51477385ef4b3114ea1534fc56209875602e962bcfa76976c45d0a045b5a2e"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.348298 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f557bc54-vbfbn" event={"ID":"d9be36cc-2136-4285-8489-93149d0136c3","Type":"ContainerStarted","Data":"c9984f0cb28c376cfa7bf27ab81afacdff61c0ec446487cec3d3f36051dd5b88"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.348890 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.348923 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.356134 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65b8b49f6-vm5h8" event={"ID":"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636","Type":"ContainerStarted","Data":"40c7d30edf14be2239634a00bcdc764f9047d1cd7b4be25002b21641947521c6"} Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.367023 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-d8m9m" podStartSLOduration=3.36700589 podStartE2EDuration="3.36700589s" podCreationTimestamp="2025-10-10 09:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:27.361337958 +0000 UTC m=+1050.377356700" watchObservedRunningTime="2025-10-10 09:28:27.36700589 +0000 UTC m=+1050.383024632" Oct 10 09:28:27 crc kubenswrapper[4669]: I1010 09:28:27.825604 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7f557bc54-vbfbn" podStartSLOduration=5.825575895 podStartE2EDuration="5.825575895s" podCreationTimestamp="2025-10-10 09:28:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:27.405458416 +0000 UTC m=+1050.421477158" watchObservedRunningTime="2025-10-10 09:28:27.825575895 +0000 UTC m=+1050.841594627" Oct 10 09:28:28 crc kubenswrapper[4669]: I1010 09:28:28.378789 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65b8b49f6-vm5h8" event={"ID":"8c9b81a2-843b-4bab-b58a-5e2a1ce2e636","Type":"ContainerStarted","Data":"b7b1fccc797bb5e42ed8ca1ff904a0333776cfbc01e8c36301917adb9ea11e0d"} Oct 10 09:28:28 crc kubenswrapper[4669]: I1010 09:28:28.379597 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:28:44 crc kubenswrapper[4669]: E1010 09:28:44.287756 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 10 09:28:44 crc kubenswrapper[4669]: E1010 09:28:44.288851 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c9rtc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(7f64ef22-59f4-471b-9726-b3673ba3ddcf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 10 09:28:44 crc kubenswrapper[4669]: E1010 09:28:44.290073 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" Oct 10 09:28:44 crc kubenswrapper[4669]: I1010 09:28:44.520914 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-central-agent" containerID="cri-o://32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49" gracePeriod=30 Oct 10 09:28:44 crc kubenswrapper[4669]: I1010 09:28:44.521095 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="sg-core" containerID="cri-o://ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8" gracePeriod=30 Oct 10 09:28:44 crc kubenswrapper[4669]: I1010 09:28:44.521192 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-notification-agent" containerID="cri-o://f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810" gracePeriod=30 Oct 10 09:28:44 crc kubenswrapper[4669]: I1010 09:28:44.549111 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-65b8b49f6-vm5h8" podStartSLOduration=18.549092218 podStartE2EDuration="18.549092218s" podCreationTimestamp="2025-10-10 09:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:28.400045115 +0000 UTC m=+1051.416063857" watchObservedRunningTime="2025-10-10 09:28:44.549092218 +0000 UTC m=+1067.565110970" Oct 10 09:28:45 crc kubenswrapper[4669]: E1010 09:28:45.376153 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 10 09:28:45 crc kubenswrapper[4669]: E1010 09:28:45.376697 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t29ld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-trw89_openstack(a1d9870d-304c-4ab1-9091-9e7e939cc6b8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 09:28:45 crc kubenswrapper[4669]: E1010 09:28:45.378119 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-trw89" podUID="a1d9870d-304c-4ab1-9091-9e7e939cc6b8" Oct 10 09:28:45 crc kubenswrapper[4669]: I1010 09:28:45.531603 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l7kr8" event={"ID":"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff","Type":"ContainerStarted","Data":"0eb010860136cb69a607f20a56d50bc6cb600cc3aa977818b62bfe45edf66223"} Oct 10 09:28:45 crc kubenswrapper[4669]: I1010 09:28:45.536814 4669 generic.go:334] "Generic (PLEG): container finished" podID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerID="ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8" exitCode=2 Oct 10 09:28:45 crc kubenswrapper[4669]: I1010 09:28:45.536867 4669 generic.go:334] "Generic (PLEG): container finished" podID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerID="32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49" exitCode=0 Oct 10 09:28:45 crc kubenswrapper[4669]: I1010 09:28:45.536927 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerDied","Data":"ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8"} Oct 10 09:28:45 crc kubenswrapper[4669]: I1010 09:28:45.536958 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerDied","Data":"32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49"} Oct 10 09:28:45 crc kubenswrapper[4669]: E1010 09:28:45.538150 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-trw89" podUID="a1d9870d-304c-4ab1-9091-9e7e939cc6b8" Oct 10 09:28:45 crc kubenswrapper[4669]: I1010 09:28:45.552741 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-l7kr8" podStartSLOduration=3.017617553 podStartE2EDuration="21.552720788s" podCreationTimestamp="2025-10-10 09:28:24 +0000 UTC" firstStartedPulling="2025-10-10 09:28:26.808519354 +0000 UTC m=+1049.824538096" lastFinishedPulling="2025-10-10 09:28:45.343622109 +0000 UTC m=+1068.359641331" observedRunningTime="2025-10-10 09:28:45.551157497 +0000 UTC m=+1068.567176239" watchObservedRunningTime="2025-10-10 09:28:45.552720788 +0000 UTC m=+1068.568739530" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.456871 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.468480 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-scripts\") pod \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.468996 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-sg-core-conf-yaml\") pod \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.469462 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-config-data\") pod \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.469610 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9rtc\" (UniqueName: \"kubernetes.io/projected/7f64ef22-59f4-471b-9726-b3673ba3ddcf-kube-api-access-c9rtc\") pod \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.469720 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-run-httpd\") pod \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.469803 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-log-httpd\") pod \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.469857 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-combined-ca-bundle\") pod \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\" (UID: \"7f64ef22-59f4-471b-9726-b3673ba3ddcf\") " Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.470230 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7f64ef22-59f4-471b-9726-b3673ba3ddcf" (UID: "7f64ef22-59f4-471b-9726-b3673ba3ddcf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.470339 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7f64ef22-59f4-471b-9726-b3673ba3ddcf" (UID: "7f64ef22-59f4-471b-9726-b3673ba3ddcf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.475765 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-scripts" (OuterVolumeSpecName: "scripts") pod "7f64ef22-59f4-471b-9726-b3673ba3ddcf" (UID: "7f64ef22-59f4-471b-9726-b3673ba3ddcf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.475760 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f64ef22-59f4-471b-9726-b3673ba3ddcf-kube-api-access-c9rtc" (OuterVolumeSpecName: "kube-api-access-c9rtc") pod "7f64ef22-59f4-471b-9726-b3673ba3ddcf" (UID: "7f64ef22-59f4-471b-9726-b3673ba3ddcf"). InnerVolumeSpecName "kube-api-access-c9rtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.500038 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7f64ef22-59f4-471b-9726-b3673ba3ddcf" (UID: "7f64ef22-59f4-471b-9726-b3673ba3ddcf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.518690 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-config-data" (OuterVolumeSpecName: "config-data") pod "7f64ef22-59f4-471b-9726-b3673ba3ddcf" (UID: "7f64ef22-59f4-471b-9726-b3673ba3ddcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.519134 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f64ef22-59f4-471b-9726-b3673ba3ddcf" (UID: "7f64ef22-59f4-471b-9726-b3673ba3ddcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.551478 4669 generic.go:334] "Generic (PLEG): container finished" podID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerID="f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810" exitCode=0 Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.551677 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerDied","Data":"f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810"} Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.551775 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.551783 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f64ef22-59f4-471b-9726-b3673ba3ddcf","Type":"ContainerDied","Data":"dfbcdbfe73307042dc35f74d8eb5581c2d3d24aac001e1724caa4abe5a37e595"} Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.551886 4669 scope.go:117] "RemoveContainer" containerID="ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.572602 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9rtc\" (UniqueName: \"kubernetes.io/projected/7f64ef22-59f4-471b-9726-b3673ba3ddcf-kube-api-access-c9rtc\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.572769 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.572855 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f64ef22-59f4-471b-9726-b3673ba3ddcf-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.572981 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.573066 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.573141 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.573191 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f64ef22-59f4-471b-9726-b3673ba3ddcf-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.579514 4669 scope.go:117] "RemoveContainer" containerID="f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.637623 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.644863 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.648722 4669 scope.go:117] "RemoveContainer" containerID="32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.656729 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:47 crc kubenswrapper[4669]: E1010 09:28:47.657080 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-notification-agent" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.657097 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-notification-agent" Oct 10 09:28:47 crc kubenswrapper[4669]: E1010 09:28:47.657125 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="sg-core" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.657132 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="sg-core" Oct 10 09:28:47 crc kubenswrapper[4669]: E1010 09:28:47.657156 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-central-agent" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.657163 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-central-agent" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.657333 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-notification-agent" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.657357 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="ceilometer-central-agent" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.657373 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" containerName="sg-core" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.659029 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.661266 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.661422 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.665598 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.676752 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ml68\" (UniqueName: \"kubernetes.io/projected/fd93c04e-6db4-4411-8390-d10cd10b8048-kube-api-access-7ml68\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.676852 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-config-data\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.676944 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-log-httpd\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.677006 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-scripts\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.677023 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.677041 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-run-httpd\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.677067 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.683187 4669 scope.go:117] "RemoveContainer" containerID="ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8" Oct 10 09:28:47 crc kubenswrapper[4669]: E1010 09:28:47.683542 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8\": container with ID starting with ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8 not found: ID does not exist" containerID="ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.683637 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8"} err="failed to get container status \"ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8\": rpc error: code = NotFound desc = could not find container \"ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8\": container with ID starting with ec784693094825949dec44ec3285bc7cd6e78758184d1cd9994a27aa1b6188b8 not found: ID does not exist" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.683682 4669 scope.go:117] "RemoveContainer" containerID="f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810" Oct 10 09:28:47 crc kubenswrapper[4669]: E1010 09:28:47.684077 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810\": container with ID starting with f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810 not found: ID does not exist" containerID="f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.684099 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810"} err="failed to get container status \"f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810\": rpc error: code = NotFound desc = could not find container \"f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810\": container with ID starting with f7b1b1cd1e7be91b680d6667f2a8ac12bb67c00eee5071b74c6938dcbfdf3810 not found: ID does not exist" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.684112 4669 scope.go:117] "RemoveContainer" containerID="32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49" Oct 10 09:28:47 crc kubenswrapper[4669]: E1010 09:28:47.684393 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49\": container with ID starting with 32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49 not found: ID does not exist" containerID="32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.684806 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49"} err="failed to get container status \"32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49\": rpc error: code = NotFound desc = could not find container \"32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49\": container with ID starting with 32cb7caf1bb0ac778260152317dc19c0f678f5f95b6123d84f2a262ec20ecb49 not found: ID does not exist" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778548 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-log-httpd\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778627 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-scripts\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778647 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778665 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-run-httpd\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778686 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778730 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ml68\" (UniqueName: \"kubernetes.io/projected/fd93c04e-6db4-4411-8390-d10cd10b8048-kube-api-access-7ml68\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778766 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-config-data\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.778946 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-log-httpd\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.779279 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-run-httpd\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.781739 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.782260 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-scripts\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.782542 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-config-data\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.783488 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.801255 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ml68\" (UniqueName: \"kubernetes.io/projected/fd93c04e-6db4-4411-8390-d10cd10b8048-kube-api-access-7ml68\") pod \"ceilometer-0\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " pod="openstack/ceilometer-0" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.804757 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f64ef22-59f4-471b-9726-b3673ba3ddcf" path="/var/lib/kubelet/pods/7f64ef22-59f4-471b-9726-b3673ba3ddcf/volumes" Oct 10 09:28:47 crc kubenswrapper[4669]: I1010 09:28:47.979239 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:28:48 crc kubenswrapper[4669]: I1010 09:28:48.440911 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:28:48 crc kubenswrapper[4669]: W1010 09:28:48.448763 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd93c04e_6db4_4411_8390_d10cd10b8048.slice/crio-a0a7f9c7a2d8fc270afbab41fccdb7505f661651003ccaf3fc5b0e250d008b74 WatchSource:0}: Error finding container a0a7f9c7a2d8fc270afbab41fccdb7505f661651003ccaf3fc5b0e250d008b74: Status 404 returned error can't find the container with id a0a7f9c7a2d8fc270afbab41fccdb7505f661651003ccaf3fc5b0e250d008b74 Oct 10 09:28:48 crc kubenswrapper[4669]: I1010 09:28:48.563221 4669 generic.go:334] "Generic (PLEG): container finished" podID="37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" containerID="0eb010860136cb69a607f20a56d50bc6cb600cc3aa977818b62bfe45edf66223" exitCode=0 Oct 10 09:28:48 crc kubenswrapper[4669]: I1010 09:28:48.563281 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l7kr8" event={"ID":"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff","Type":"ContainerDied","Data":"0eb010860136cb69a607f20a56d50bc6cb600cc3aa977818b62bfe45edf66223"} Oct 10 09:28:48 crc kubenswrapper[4669]: I1010 09:28:48.571885 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerStarted","Data":"a0a7f9c7a2d8fc270afbab41fccdb7505f661651003ccaf3fc5b0e250d008b74"} Oct 10 09:28:49 crc kubenswrapper[4669]: I1010 09:28:49.581559 4669 generic.go:334] "Generic (PLEG): container finished" podID="50102622-b6c0-4fe9-8eab-b4027da1a36a" containerID="416e34c6fab75954a71767e78290564c70e9e1feae973fab44258c9fa061165f" exitCode=0 Oct 10 09:28:49 crc kubenswrapper[4669]: I1010 09:28:49.581707 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d8m9m" event={"ID":"50102622-b6c0-4fe9-8eab-b4027da1a36a","Type":"ContainerDied","Data":"416e34c6fab75954a71767e78290564c70e9e1feae973fab44258c9fa061165f"} Oct 10 09:28:49 crc kubenswrapper[4669]: I1010 09:28:49.585266 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerStarted","Data":"6d961d1a9ef9537910a8a6e12d37020feadbdd40bba00d47c30ca513cc2e3e9e"} Oct 10 09:28:49 crc kubenswrapper[4669]: I1010 09:28:49.955872 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.123224 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkrc6\" (UniqueName: \"kubernetes.io/projected/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-kube-api-access-tkrc6\") pod \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.123351 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-db-sync-config-data\") pod \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.123546 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-combined-ca-bundle\") pod \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\" (UID: \"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff\") " Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.129185 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-kube-api-access-tkrc6" (OuterVolumeSpecName: "kube-api-access-tkrc6") pod "37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" (UID: "37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff"). InnerVolumeSpecName "kube-api-access-tkrc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.129741 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" (UID: "37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.148608 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" (UID: "37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.227354 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkrc6\" (UniqueName: \"kubernetes.io/projected/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-kube-api-access-tkrc6\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.227410 4669 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.227430 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.611794 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerStarted","Data":"2559692f8cbd2848abecd8a581a5fcee60d267ffaba17ae2aa3123d36e97cec3"} Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.612261 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerStarted","Data":"403adf74fc07ea0cf34dbcebe137f53a8ba5e31d91bf573e559672303cb82066"} Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.620535 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-l7kr8" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.621681 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-l7kr8" event={"ID":"37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff","Type":"ContainerDied","Data":"30b5cfbab5e9eefe0de80cab4bbe44acb0615a2c66884da7393b2afc60642333"} Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.621716 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30b5cfbab5e9eefe0de80cab4bbe44acb0615a2c66884da7393b2afc60642333" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.912258 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-686db9dfc7-hg7z4"] Oct 10 09:28:50 crc kubenswrapper[4669]: E1010 09:28:50.913647 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" containerName="barbican-db-sync" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.913667 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" containerName="barbican-db-sync" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.913971 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" containerName="barbican-db-sync" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.916472 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.924514 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.938970 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.939115 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-j5xr7" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.960647 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-686db9dfc7-hg7z4"] Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.981789 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p"] Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.983193 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:50 crc kubenswrapper[4669]: I1010 09:28:50.989006 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.031169 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p"] Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.052409 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-combined-ca-bundle\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.052450 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-config-data\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.052469 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-config-data-custom\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.052518 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtfl5\" (UniqueName: \"kubernetes.io/projected/58f4d796-043f-48f0-a649-99fb297c694a-kube-api-access-qtfl5\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.052573 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58f4d796-043f-48f0-a649-99fb297c694a-logs\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.054033 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65d5b6b857-c2476"] Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.056394 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.066256 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65d5b6b857-c2476"] Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.146937 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-cbc79ddd8-vqntz"] Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.148237 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.153436 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155354 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/11452488-a27a-4408-97d5-4e04eb598cad-kube-api-access-g5bxq\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155388 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-config-data-custom\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155417 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58f4d796-043f-48f0-a649-99fb297c694a-logs\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155450 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-combined-ca-bundle\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155467 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-nb\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155496 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-config\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155527 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-combined-ca-bundle\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155545 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-config-data\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155564 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-config-data-custom\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155598 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-config-data\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155631 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-dns-svc\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155650 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtfl5\" (UniqueName: \"kubernetes.io/projected/58f4d796-043f-48f0-a649-99fb297c694a-kube-api-access-qtfl5\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155683 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb7dc377-170c-4504-a61b-386110e80526-logs\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155704 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-sb\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.155722 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnq5k\" (UniqueName: \"kubernetes.io/projected/bb7dc377-170c-4504-a61b-386110e80526-kube-api-access-rnq5k\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.156360 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58f4d796-043f-48f0-a649-99fb297c694a-logs\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.165908 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-combined-ca-bundle\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.179404 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-config-data\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.182491 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cbc79ddd8-vqntz"] Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.188724 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58f4d796-043f-48f0-a649-99fb297c694a-config-data-custom\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.219292 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtfl5\" (UniqueName: \"kubernetes.io/projected/58f4d796-043f-48f0-a649-99fb297c694a-kube-api-access-qtfl5\") pod \"barbican-worker-686db9dfc7-hg7z4\" (UID: \"58f4d796-043f-48f0-a649-99fb297c694a\") " pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.249829 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.254206 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-686db9dfc7-hg7z4" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256769 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-dns-svc\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256803 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06094d4a-33e6-4b9f-b291-90262283b33a-logs\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256834 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb7dc377-170c-4504-a61b-386110e80526-logs\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256858 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-sb\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256877 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnq5k\" (UniqueName: \"kubernetes.io/projected/bb7dc377-170c-4504-a61b-386110e80526-kube-api-access-rnq5k\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256909 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/11452488-a27a-4408-97d5-4e04eb598cad-kube-api-access-g5bxq\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256929 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-config-data-custom\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256949 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data-custom\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256968 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-combined-ca-bundle\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.256993 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-combined-ca-bundle\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.257009 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-nb\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.257030 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6dnj\" (UniqueName: \"kubernetes.io/projected/06094d4a-33e6-4b9f-b291-90262283b33a-kube-api-access-w6dnj\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.257048 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-config\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.257079 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.257104 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-config-data\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.257599 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-dns-svc\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.258087 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb7dc377-170c-4504-a61b-386110e80526-logs\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.258443 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-nb\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.259956 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-sb\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.263284 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-combined-ca-bundle\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.265047 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-config\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.272065 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-config-data-custom\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.292480 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7dc377-170c-4504-a61b-386110e80526-config-data\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.296651 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnq5k\" (UniqueName: \"kubernetes.io/projected/bb7dc377-170c-4504-a61b-386110e80526-kube-api-access-rnq5k\") pod \"barbican-keystone-listener-5fcbcfb6bb-9hc2p\" (UID: \"bb7dc377-170c-4504-a61b-386110e80526\") " pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.299114 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/11452488-a27a-4408-97d5-4e04eb598cad-kube-api-access-g5bxq\") pod \"dnsmasq-dns-65d5b6b857-c2476\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.319862 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.358916 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86smk\" (UniqueName: \"kubernetes.io/projected/50102622-b6c0-4fe9-8eab-b4027da1a36a-kube-api-access-86smk\") pod \"50102622-b6c0-4fe9-8eab-b4027da1a36a\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.359305 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-config\") pod \"50102622-b6c0-4fe9-8eab-b4027da1a36a\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.359331 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-combined-ca-bundle\") pod \"50102622-b6c0-4fe9-8eab-b4027da1a36a\" (UID: \"50102622-b6c0-4fe9-8eab-b4027da1a36a\") " Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.360388 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data-custom\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.360525 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-combined-ca-bundle\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.360577 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6dnj\" (UniqueName: \"kubernetes.io/projected/06094d4a-33e6-4b9f-b291-90262283b33a-kube-api-access-w6dnj\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.361124 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.361302 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06094d4a-33e6-4b9f-b291-90262283b33a-logs\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.366137 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50102622-b6c0-4fe9-8eab-b4027da1a36a-kube-api-access-86smk" (OuterVolumeSpecName: "kube-api-access-86smk") pod "50102622-b6c0-4fe9-8eab-b4027da1a36a" (UID: "50102622-b6c0-4fe9-8eab-b4027da1a36a"). InnerVolumeSpecName "kube-api-access-86smk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.371372 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06094d4a-33e6-4b9f-b291-90262283b33a-logs\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.373123 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data-custom\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.375602 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.376815 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-combined-ca-bundle\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.377450 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.397546 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6dnj\" (UniqueName: \"kubernetes.io/projected/06094d4a-33e6-4b9f-b291-90262283b33a-kube-api-access-w6dnj\") pod \"barbican-api-cbc79ddd8-vqntz\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.410001 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50102622-b6c0-4fe9-8eab-b4027da1a36a" (UID: "50102622-b6c0-4fe9-8eab-b4027da1a36a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.418668 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-config" (OuterVolumeSpecName: "config") pod "50102622-b6c0-4fe9-8eab-b4027da1a36a" (UID: "50102622-b6c0-4fe9-8eab-b4027da1a36a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.463055 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86smk\" (UniqueName: \"kubernetes.io/projected/50102622-b6c0-4fe9-8eab-b4027da1a36a-kube-api-access-86smk\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.463085 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.463095 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50102622-b6c0-4fe9-8eab-b4027da1a36a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.564814 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.657253 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d8m9m" event={"ID":"50102622-b6c0-4fe9-8eab-b4027da1a36a","Type":"ContainerDied","Data":"700baef58b6a24e63d6fc452bab60dc8f0730748c5d3e29b2854aea725ffce78"} Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.657522 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="700baef58b6a24e63d6fc452bab60dc8f0730748c5d3e29b2854aea725ffce78" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.657294 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d8m9m" Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.758432 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-686db9dfc7-hg7z4"] Oct 10 09:28:51 crc kubenswrapper[4669]: W1010 09:28:51.769284 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58f4d796_043f_48f0_a649_99fb297c694a.slice/crio-c8474f41b158b5ad202b4f2a241f9457d1cf8a05be99f6d77a94db4c2dd64748 WatchSource:0}: Error finding container c8474f41b158b5ad202b4f2a241f9457d1cf8a05be99f6d77a94db4c2dd64748: Status 404 returned error can't find the container with id c8474f41b158b5ad202b4f2a241f9457d1cf8a05be99f6d77a94db4c2dd64748 Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.879518 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p"] Oct 10 09:28:51 crc kubenswrapper[4669]: I1010 09:28:51.906027 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d5b6b857-c2476"] Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.006139 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mgpjg"] Oct 10 09:28:52 crc kubenswrapper[4669]: E1010 09:28:52.006632 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50102622-b6c0-4fe9-8eab-b4027da1a36a" containerName="neutron-db-sync" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.006646 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="50102622-b6c0-4fe9-8eab-b4027da1a36a" containerName="neutron-db-sync" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.006868 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="50102622-b6c0-4fe9-8eab-b4027da1a36a" containerName="neutron-db-sync" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.007890 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.017537 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mgpjg"] Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.060994 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d5b6b857-c2476"] Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.081209 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.081544 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-config\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.083758 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.083814 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-dns-svc\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.084058 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm677\" (UniqueName: \"kubernetes.io/projected/eb84008d-41fb-46a0-97d9-4237d4575fe6-kube-api-access-wm677\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.096374 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-588b8ffdfb-2xrm4"] Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.101499 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.107992 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.112742 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.112875 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9zprf" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.113030 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.133043 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-588b8ffdfb-2xrm4"] Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.189852 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cbc79ddd8-vqntz"] Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.219669 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.219731 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-config\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.219769 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-dns-svc\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.219851 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-combined-ca-bundle\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.219893 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm677\" (UniqueName: \"kubernetes.io/projected/eb84008d-41fb-46a0-97d9-4237d4575fe6-kube-api-access-wm677\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.219972 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hgfm\" (UniqueName: \"kubernetes.io/projected/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-kube-api-access-9hgfm\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.220026 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.220080 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-httpd-config\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.220144 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-config\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.220218 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-ovndb-tls-certs\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.221935 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.222510 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-config\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.223102 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-dns-svc\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.250731 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.256083 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm677\" (UniqueName: \"kubernetes.io/projected/eb84008d-41fb-46a0-97d9-4237d4575fe6-kube-api-access-wm677\") pod \"dnsmasq-dns-6dc8d75dbf-mgpjg\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.323237 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hgfm\" (UniqueName: \"kubernetes.io/projected/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-kube-api-access-9hgfm\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.323314 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-httpd-config\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.323379 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-ovndb-tls-certs\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.323412 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-config\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.323451 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-combined-ca-bundle\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.332360 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-config\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.333268 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-combined-ca-bundle\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.334031 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-ovndb-tls-certs\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.344702 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-httpd-config\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.349984 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hgfm\" (UniqueName: \"kubernetes.io/projected/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-kube-api-access-9hgfm\") pod \"neutron-588b8ffdfb-2xrm4\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.422394 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.465134 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.678292 4669 generic.go:334] "Generic (PLEG): container finished" podID="11452488-a27a-4408-97d5-4e04eb598cad" containerID="3d678a6d538f73890e5a282e671c2b273d5310c726d928364d686501972df023" exitCode=0 Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.678367 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d5b6b857-c2476" event={"ID":"11452488-a27a-4408-97d5-4e04eb598cad","Type":"ContainerDied","Data":"3d678a6d538f73890e5a282e671c2b273d5310c726d928364d686501972df023"} Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.678398 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d5b6b857-c2476" event={"ID":"11452488-a27a-4408-97d5-4e04eb598cad","Type":"ContainerStarted","Data":"778b5f21eb943d48366df5704e36ec54b9d39965da737c66032feae26c35e547"} Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.782229 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-686db9dfc7-hg7z4" event={"ID":"58f4d796-043f-48f0-a649-99fb297c694a","Type":"ContainerStarted","Data":"c8474f41b158b5ad202b4f2a241f9457d1cf8a05be99f6d77a94db4c2dd64748"} Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.790308 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" event={"ID":"bb7dc377-170c-4504-a61b-386110e80526","Type":"ContainerStarted","Data":"f2639dfccca548ebb8d74156709f8f52da2cad218aad3c7a41ed2af70733e54d"} Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.815822 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerStarted","Data":"68a72109ef57ed0c00c70dcfea6e5b064bb1f0728abc24e418d154464647f1c8"} Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.816768 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.824981 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbc79ddd8-vqntz" event={"ID":"06094d4a-33e6-4b9f-b291-90262283b33a","Type":"ContainerStarted","Data":"3768ca9b47d1a830f88af92ef8cd83104e84dfc9021dd33f4521c7ce0a0f7460"} Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.825031 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbc79ddd8-vqntz" event={"ID":"06094d4a-33e6-4b9f-b291-90262283b33a","Type":"ContainerStarted","Data":"84666ba9a358ae2048259324909f066d5fcd89d8776f8cf9c597c315a6331293"} Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.849688 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.849746 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.861408 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.769982399 podStartE2EDuration="5.861387535s" podCreationTimestamp="2025-10-10 09:28:47 +0000 UTC" firstStartedPulling="2025-10-10 09:28:48.454316693 +0000 UTC m=+1071.470335435" lastFinishedPulling="2025-10-10 09:28:51.545721829 +0000 UTC m=+1074.561740571" observedRunningTime="2025-10-10 09:28:52.850477515 +0000 UTC m=+1075.866496257" watchObservedRunningTime="2025-10-10 09:28:52.861387535 +0000 UTC m=+1075.877406277" Oct 10 09:28:52 crc kubenswrapper[4669]: I1010 09:28:52.892093 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-cbc79ddd8-vqntz" podStartSLOduration=1.8920766709999999 podStartE2EDuration="1.892076671s" podCreationTimestamp="2025-10-10 09:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:52.891544834 +0000 UTC m=+1075.907563566" watchObservedRunningTime="2025-10-10 09:28:52.892076671 +0000 UTC m=+1075.908095413" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.060012 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mgpjg"] Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.241854 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.289576 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-config\") pod \"11452488-a27a-4408-97d5-4e04eb598cad\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.289769 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-dns-svc\") pod \"11452488-a27a-4408-97d5-4e04eb598cad\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.289895 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/11452488-a27a-4408-97d5-4e04eb598cad-kube-api-access-g5bxq\") pod \"11452488-a27a-4408-97d5-4e04eb598cad\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.289936 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-nb\") pod \"11452488-a27a-4408-97d5-4e04eb598cad\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.289962 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-sb\") pod \"11452488-a27a-4408-97d5-4e04eb598cad\" (UID: \"11452488-a27a-4408-97d5-4e04eb598cad\") " Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.305497 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11452488-a27a-4408-97d5-4e04eb598cad-kube-api-access-g5bxq" (OuterVolumeSpecName: "kube-api-access-g5bxq") pod "11452488-a27a-4408-97d5-4e04eb598cad" (UID: "11452488-a27a-4408-97d5-4e04eb598cad"). InnerVolumeSpecName "kube-api-access-g5bxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.340667 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11452488-a27a-4408-97d5-4e04eb598cad" (UID: "11452488-a27a-4408-97d5-4e04eb598cad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.396253 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11452488-a27a-4408-97d5-4e04eb598cad" (UID: "11452488-a27a-4408-97d5-4e04eb598cad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.397235 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5bxq\" (UniqueName: \"kubernetes.io/projected/11452488-a27a-4408-97d5-4e04eb598cad-kube-api-access-g5bxq\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.397257 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.397266 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.408730 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-588b8ffdfb-2xrm4"] Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.428195 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-config" (OuterVolumeSpecName: "config") pod "11452488-a27a-4408-97d5-4e04eb598cad" (UID: "11452488-a27a-4408-97d5-4e04eb598cad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.428571 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11452488-a27a-4408-97d5-4e04eb598cad" (UID: "11452488-a27a-4408-97d5-4e04eb598cad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.506464 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.506753 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11452488-a27a-4408-97d5-4e04eb598cad-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.836486 4669 generic.go:334] "Generic (PLEG): container finished" podID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerID="0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4" exitCode=0 Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.836553 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" event={"ID":"eb84008d-41fb-46a0-97d9-4237d4575fe6","Type":"ContainerDied","Data":"0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4"} Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.836579 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" event={"ID":"eb84008d-41fb-46a0-97d9-4237d4575fe6","Type":"ContainerStarted","Data":"bdcc73d3b4c24190017019770ddb3d2a2174085d33bb17c53969a736245b1c74"} Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.840113 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbc79ddd8-vqntz" event={"ID":"06094d4a-33e6-4b9f-b291-90262283b33a","Type":"ContainerStarted","Data":"c398199fc54e6e39ff3b3fa0e9e4e7d32fa49777018d2f46a5cbcb3c9cdd05d1"} Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.850138 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-588b8ffdfb-2xrm4" event={"ID":"c92807a3-b0ae-48c0-ba7b-1e341aabde2a","Type":"ContainerStarted","Data":"1ea8850ed5c4772058ca1f9fe9b9d04a3cb3fb182cd1cfd9bd6c6673f5f00026"} Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.850176 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-588b8ffdfb-2xrm4" event={"ID":"c92807a3-b0ae-48c0-ba7b-1e341aabde2a","Type":"ContainerStarted","Data":"917abb5150a2a61d14e59923222d687d09ba5a7b71d3ae749faf65471d973843"} Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.880374 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d5b6b857-c2476" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.881047 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d5b6b857-c2476" event={"ID":"11452488-a27a-4408-97d5-4e04eb598cad","Type":"ContainerDied","Data":"778b5f21eb943d48366df5704e36ec54b9d39965da737c66032feae26c35e547"} Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.881090 4669 scope.go:117] "RemoveContainer" containerID="3d678a6d538f73890e5a282e671c2b273d5310c726d928364d686501972df023" Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.934739 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d5b6b857-c2476"] Oct 10 09:28:53 crc kubenswrapper[4669]: I1010 09:28:53.937436 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65d5b6b857-c2476"] Oct 10 09:28:54 crc kubenswrapper[4669]: I1010 09:28:54.940729 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.169545 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7f557bc54-vbfbn" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.825931 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11452488-a27a-4408-97d5-4e04eb598cad" path="/var/lib/kubelet/pods/11452488-a27a-4408-97d5-4e04eb598cad/volumes" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.907834 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" event={"ID":"bb7dc377-170c-4504-a61b-386110e80526","Type":"ContainerStarted","Data":"002455b0396a48e3a5c13f561cecf61b1db9ec2def70c4f2d6ec141a1fdfa97e"} Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.910822 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-588b8ffdfb-2xrm4" event={"ID":"c92807a3-b0ae-48c0-ba7b-1e341aabde2a","Type":"ContainerStarted","Data":"272a322b633e4955e9bedb2be9f7cbb125c4b0428dd4109d2d1ab55e79bdc15a"} Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.911792 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.918239 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" event={"ID":"eb84008d-41fb-46a0-97d9-4237d4575fe6","Type":"ContainerStarted","Data":"ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec"} Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.919173 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.923994 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-686db9dfc7-hg7z4" event={"ID":"58f4d796-043f-48f0-a649-99fb297c694a","Type":"ContainerStarted","Data":"3a178e87f1f2cfe61efd102f8d6f39cd7f1d390a77d926d09b4230db84c4dfeb"} Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.932892 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6d556ffc9f-2x4qt"] Oct 10 09:28:55 crc kubenswrapper[4669]: E1010 09:28:55.933228 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11452488-a27a-4408-97d5-4e04eb598cad" containerName="init" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.933243 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="11452488-a27a-4408-97d5-4e04eb598cad" containerName="init" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.933399 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="11452488-a27a-4408-97d5-4e04eb598cad" containerName="init" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.939418 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-588b8ffdfb-2xrm4" podStartSLOduration=3.93940306 podStartE2EDuration="3.93940306s" podCreationTimestamp="2025-10-10 09:28:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:55.935291118 +0000 UTC m=+1078.951309860" watchObservedRunningTime="2025-10-10 09:28:55.93940306 +0000 UTC m=+1078.955421802" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.944098 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.947285 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.965221 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.967850 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6d556ffc9f-2x4qt"] Oct 10 09:28:55 crc kubenswrapper[4669]: I1010 09:28:55.977981 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" podStartSLOduration=4.977957969 podStartE2EDuration="4.977957969s" podCreationTimestamp="2025-10-10 09:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:55.976931876 +0000 UTC m=+1078.992950618" watchObservedRunningTime="2025-10-10 09:28:55.977957969 +0000 UTC m=+1078.993976711" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.090440 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-httpd-config\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.090512 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxfnb\" (UniqueName: \"kubernetes.io/projected/5bf9895c-806b-4a83-ae94-7530013432fb-kube-api-access-jxfnb\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.090542 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-config\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.090568 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-internal-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.090603 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-combined-ca-bundle\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.090663 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-public-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.090681 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-ovndb-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.192872 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-httpd-config\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.193321 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxfnb\" (UniqueName: \"kubernetes.io/projected/5bf9895c-806b-4a83-ae94-7530013432fb-kube-api-access-jxfnb\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.193356 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-config\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.193382 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-internal-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.193409 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-combined-ca-bundle\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.193472 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-public-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.193506 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-ovndb-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.200383 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-combined-ca-bundle\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.202878 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-ovndb-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.203328 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-httpd-config\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.205391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-internal-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.206405 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-config\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.206892 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bf9895c-806b-4a83-ae94-7530013432fb-public-tls-certs\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.214321 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxfnb\" (UniqueName: \"kubernetes.io/projected/5bf9895c-806b-4a83-ae94-7530013432fb-kube-api-access-jxfnb\") pod \"neutron-6d556ffc9f-2x4qt\" (UID: \"5bf9895c-806b-4a83-ae94-7530013432fb\") " pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.260154 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.839417 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6d556ffc9f-2x4qt"] Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.934441 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d556ffc9f-2x4qt" event={"ID":"5bf9895c-806b-4a83-ae94-7530013432fb","Type":"ContainerStarted","Data":"defc497d7b103af7113b6a09a9d86790706fb4f52a46dbc0321533de92ad5421"} Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.936395 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-686db9dfc7-hg7z4" event={"ID":"58f4d796-043f-48f0-a649-99fb297c694a","Type":"ContainerStarted","Data":"5ab419bf6d7361608897c31c1fc3f86198017a4ac959beaa45302d991a2ed218"} Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.950095 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" event={"ID":"bb7dc377-170c-4504-a61b-386110e80526","Type":"ContainerStarted","Data":"00c69027d0818ab364ec42242f7deb749fa209b89fdd2dd0e8d91ea76a9c27ab"} Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.969855 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-686db9dfc7-hg7z4" podStartSLOduration=3.460246939 podStartE2EDuration="6.969832231s" podCreationTimestamp="2025-10-10 09:28:50 +0000 UTC" firstStartedPulling="2025-10-10 09:28:51.793238953 +0000 UTC m=+1074.809257695" lastFinishedPulling="2025-10-10 09:28:55.302824245 +0000 UTC m=+1078.318842987" observedRunningTime="2025-10-10 09:28:56.965481681 +0000 UTC m=+1079.981500423" watchObservedRunningTime="2025-10-10 09:28:56.969832231 +0000 UTC m=+1079.985850973" Oct 10 09:28:56 crc kubenswrapper[4669]: I1010 09:28:56.988807 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5fcbcfb6bb-9hc2p" podStartSLOduration=3.6448936610000002 podStartE2EDuration="6.98878977s" podCreationTimestamp="2025-10-10 09:28:50 +0000 UTC" firstStartedPulling="2025-10-10 09:28:51.894906479 +0000 UTC m=+1074.910925211" lastFinishedPulling="2025-10-10 09:28:55.238802588 +0000 UTC m=+1078.254821320" observedRunningTime="2025-10-10 09:28:56.986820397 +0000 UTC m=+1080.002839139" watchObservedRunningTime="2025-10-10 09:28:56.98878977 +0000 UTC m=+1080.004808512" Oct 10 09:28:57 crc kubenswrapper[4669]: I1010 09:28:57.958293 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d556ffc9f-2x4qt" event={"ID":"5bf9895c-806b-4a83-ae94-7530013432fb","Type":"ContainerStarted","Data":"cb63eac8139278b0b0104b0fcef241b0e6fb8447171c919bd047d68cbde8f794"} Oct 10 09:28:57 crc kubenswrapper[4669]: I1010 09:28:57.958674 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d556ffc9f-2x4qt" event={"ID":"5bf9895c-806b-4a83-ae94-7530013432fb","Type":"ContainerStarted","Data":"22137bc275b982a714e1551add4c01d4930a60b3bb385bb2deadeb96a9f56b0b"} Oct 10 09:28:57 crc kubenswrapper[4669]: I1010 09:28:57.959772 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.116419 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6d556ffc9f-2x4qt" podStartSLOduration=3.116400403 podStartE2EDuration="3.116400403s" podCreationTimestamp="2025-10-10 09:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:28:58.002490023 +0000 UTC m=+1081.018508775" watchObservedRunningTime="2025-10-10 09:28:58.116400403 +0000 UTC m=+1081.132419145" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.117821 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8c86b488-lgxwv"] Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.119346 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.126748 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.132741 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.143757 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8c86b488-lgxwv"] Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.236278 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-logs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.236322 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tnv6\" (UniqueName: \"kubernetes.io/projected/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-kube-api-access-9tnv6\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.236418 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-public-tls-certs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.236463 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-combined-ca-bundle\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.236484 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-config-data-custom\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.236504 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-config-data\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.236790 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-internal-tls-certs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.337979 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-combined-ca-bundle\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.338563 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-config-data-custom\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.338663 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-config-data\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.338768 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-internal-tls-certs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.338894 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-logs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.338986 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tnv6\" (UniqueName: \"kubernetes.io/projected/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-kube-api-access-9tnv6\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.339120 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-public-tls-certs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.339299 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-logs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.345309 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-config-data-custom\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.347879 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-combined-ca-bundle\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.350184 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-config-data\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.350543 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-public-tls-certs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.356853 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-internal-tls-certs\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.359971 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tnv6\" (UniqueName: \"kubernetes.io/projected/9306ad2c-0c67-4a1a-8031-19d4cddce8d0-kube-api-access-9tnv6\") pod \"barbican-api-8c86b488-lgxwv\" (UID: \"9306ad2c-0c67-4a1a-8031-19d4cddce8d0\") " pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.432988 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:28:58 crc kubenswrapper[4669]: W1010 09:28:58.957742 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9306ad2c_0c67_4a1a_8031_19d4cddce8d0.slice/crio-51dc17abdaad1f70fbd95942ed31ea52a62a6c8471f0d65c7ae54a5af1593d1d WatchSource:0}: Error finding container 51dc17abdaad1f70fbd95942ed31ea52a62a6c8471f0d65c7ae54a5af1593d1d: Status 404 returned error can't find the container with id 51dc17abdaad1f70fbd95942ed31ea52a62a6c8471f0d65c7ae54a5af1593d1d Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.967662 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8c86b488-lgxwv"] Oct 10 09:28:58 crc kubenswrapper[4669]: I1010 09:28:58.982710 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8c86b488-lgxwv" event={"ID":"9306ad2c-0c67-4a1a-8031-19d4cddce8d0","Type":"ContainerStarted","Data":"51dc17abdaad1f70fbd95942ed31ea52a62a6c8471f0d65c7ae54a5af1593d1d"} Oct 10 09:28:59 crc kubenswrapper[4669]: I1010 09:28:59.993644 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8c86b488-lgxwv" event={"ID":"9306ad2c-0c67-4a1a-8031-19d4cddce8d0","Type":"ContainerStarted","Data":"9ecc0fc76e5454ddf4ee010d17f4dc0384dce82ce635d403e9d5e6594948860c"} Oct 10 09:29:00 crc kubenswrapper[4669]: I1010 09:29:00.313071 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-65b8b49f6-vm5h8" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.241961 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.243664 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.254186 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.254454 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bdzz7" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.256305 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.261549 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.311850 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2bb8c92a-28c8-4424-94c2-c577f0b6b486-openstack-config-secret\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.312105 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb8c92a-28c8-4424-94c2-c577f0b6b486-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.312232 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2bb8c92a-28c8-4424-94c2-c577f0b6b486-openstack-config\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.312372 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s4st\" (UniqueName: \"kubernetes.io/projected/2bb8c92a-28c8-4424-94c2-c577f0b6b486-kube-api-access-7s4st\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.414062 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2bb8c92a-28c8-4424-94c2-c577f0b6b486-openstack-config\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.414209 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s4st\" (UniqueName: \"kubernetes.io/projected/2bb8c92a-28c8-4424-94c2-c577f0b6b486-kube-api-access-7s4st\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.414358 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2bb8c92a-28c8-4424-94c2-c577f0b6b486-openstack-config-secret\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.414415 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb8c92a-28c8-4424-94c2-c577f0b6b486-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.414980 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2bb8c92a-28c8-4424-94c2-c577f0b6b486-openstack-config\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.421330 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2bb8c92a-28c8-4424-94c2-c577f0b6b486-openstack-config-secret\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.421485 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bb8c92a-28c8-4424-94c2-c577f0b6b486-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.424256 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.437709 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s4st\" (UniqueName: \"kubernetes.io/projected/2bb8c92a-28c8-4424-94c2-c577f0b6b486-kube-api-access-7s4st\") pod \"openstackclient\" (UID: \"2bb8c92a-28c8-4424-94c2-c577f0b6b486\") " pod="openstack/openstackclient" Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.496327 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-mcjwv"] Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.496759 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" podUID="60f556bd-2a52-4976-8709-0d79392c793e" containerName="dnsmasq-dns" containerID="cri-o://054286cb2a84d97a0999fb90f43bed134302206b4956068b9a4b7e2292dd530a" gracePeriod=10 Oct 10 09:29:02 crc kubenswrapper[4669]: I1010 09:29:02.563707 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.037461 4669 generic.go:334] "Generic (PLEG): container finished" podID="60f556bd-2a52-4976-8709-0d79392c793e" containerID="054286cb2a84d97a0999fb90f43bed134302206b4956068b9a4b7e2292dd530a" exitCode=0 Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.038061 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" event={"ID":"60f556bd-2a52-4976-8709-0d79392c793e","Type":"ContainerDied","Data":"054286cb2a84d97a0999fb90f43bed134302206b4956068b9a4b7e2292dd530a"} Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.041119 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8c86b488-lgxwv" event={"ID":"9306ad2c-0c67-4a1a-8031-19d4cddce8d0","Type":"ContainerStarted","Data":"b9d9010c348d95f1c992b7343aa1438215157d6c68910e1dafc7209c40734198"} Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.041691 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.041743 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.074422 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8c86b488-lgxwv" podStartSLOduration=5.074399467 podStartE2EDuration="5.074399467s" podCreationTimestamp="2025-10-10 09:28:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:03.066362359 +0000 UTC m=+1086.082381101" watchObservedRunningTime="2025-10-10 09:29:03.074399467 +0000 UTC m=+1086.090418209" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.110029 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.242303 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-dns-svc\") pod \"60f556bd-2a52-4976-8709-0d79392c793e\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.242454 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqx4b\" (UniqueName: \"kubernetes.io/projected/60f556bd-2a52-4976-8709-0d79392c793e-kube-api-access-jqx4b\") pod \"60f556bd-2a52-4976-8709-0d79392c793e\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.242487 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-config\") pod \"60f556bd-2a52-4976-8709-0d79392c793e\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.242593 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-nb\") pod \"60f556bd-2a52-4976-8709-0d79392c793e\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.242609 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-sb\") pod \"60f556bd-2a52-4976-8709-0d79392c793e\" (UID: \"60f556bd-2a52-4976-8709-0d79392c793e\") " Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.253722 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60f556bd-2a52-4976-8709-0d79392c793e-kube-api-access-jqx4b" (OuterVolumeSpecName: "kube-api-access-jqx4b") pod "60f556bd-2a52-4976-8709-0d79392c793e" (UID: "60f556bd-2a52-4976-8709-0d79392c793e"). InnerVolumeSpecName "kube-api-access-jqx4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.283463 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.334167 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60f556bd-2a52-4976-8709-0d79392c793e" (UID: "60f556bd-2a52-4976-8709-0d79392c793e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.345756 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.345787 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqx4b\" (UniqueName: \"kubernetes.io/projected/60f556bd-2a52-4976-8709-0d79392c793e-kube-api-access-jqx4b\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.352990 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60f556bd-2a52-4976-8709-0d79392c793e" (UID: "60f556bd-2a52-4976-8709-0d79392c793e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.358656 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "60f556bd-2a52-4976-8709-0d79392c793e" (UID: "60f556bd-2a52-4976-8709-0d79392c793e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.377564 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-config" (OuterVolumeSpecName: "config") pod "60f556bd-2a52-4976-8709-0d79392c793e" (UID: "60f556bd-2a52-4976-8709-0d79392c793e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.447204 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.447236 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:03 crc kubenswrapper[4669]: I1010 09:29:03.447245 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60f556bd-2a52-4976-8709-0d79392c793e-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.057818 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" event={"ID":"60f556bd-2a52-4976-8709-0d79392c793e","Type":"ContainerDied","Data":"6d295bbf5add68958b27cea84353a0d9dc833780f5002ed66cc82f54b341d7e9"} Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.058430 4669 scope.go:117] "RemoveContainer" containerID="054286cb2a84d97a0999fb90f43bed134302206b4956068b9a4b7e2292dd530a" Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.058578 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-mcjwv" Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.064292 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-trw89" event={"ID":"a1d9870d-304c-4ab1-9091-9e7e939cc6b8","Type":"ContainerStarted","Data":"4e6e1e2d9fa7339fbcfd04e37c884a1eb4d3cbfc91ed126a8acf41e26a14a3dc"} Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.068914 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2bb8c92a-28c8-4424-94c2-c577f0b6b486","Type":"ContainerStarted","Data":"169e462ef1cb0d40a5eef8a89d7a6b67c8221499b00b645809a8fe3ff756a1ed"} Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.090199 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-mcjwv"] Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.104660 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-mcjwv"] Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.105845 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-trw89" podStartSLOduration=5.293653417 podStartE2EDuration="41.105829411s" podCreationTimestamp="2025-10-10 09:28:23 +0000 UTC" firstStartedPulling="2025-10-10 09:28:26.836891616 +0000 UTC m=+1049.852910358" lastFinishedPulling="2025-10-10 09:29:02.64906761 +0000 UTC m=+1085.665086352" observedRunningTime="2025-10-10 09:29:04.105574062 +0000 UTC m=+1087.121592804" watchObservedRunningTime="2025-10-10 09:29:04.105829411 +0000 UTC m=+1087.121848143" Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.124571 4669 scope.go:117] "RemoveContainer" containerID="4561a054c992e11cf2817733da92c62f59bd37964a58898b0025e6449112fd32" Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.705019 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:29:04 crc kubenswrapper[4669]: I1010 09:29:04.806320 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:29:05 crc kubenswrapper[4669]: I1010 09:29:05.805346 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60f556bd-2a52-4976-8709-0d79392c793e" path="/var/lib/kubelet/pods/60f556bd-2a52-4976-8709-0d79392c793e/volumes" Oct 10 09:29:06 crc kubenswrapper[4669]: I1010 09:29:06.672218 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:29:10 crc kubenswrapper[4669]: I1010 09:29:10.850274 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8c86b488-lgxwv" Oct 10 09:29:10 crc kubenswrapper[4669]: I1010 09:29:10.909911 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-cbc79ddd8-vqntz"] Oct 10 09:29:10 crc kubenswrapper[4669]: I1010 09:29:10.910182 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-cbc79ddd8-vqntz" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api-log" containerID="cri-o://3768ca9b47d1a830f88af92ef8cd83104e84dfc9021dd33f4521c7ce0a0f7460" gracePeriod=30 Oct 10 09:29:10 crc kubenswrapper[4669]: I1010 09:29:10.910472 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-cbc79ddd8-vqntz" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api" containerID="cri-o://c398199fc54e6e39ff3b3fa0e9e4e7d32fa49777018d2f46a5cbcb3c9cdd05d1" gracePeriod=30 Oct 10 09:29:11 crc kubenswrapper[4669]: I1010 09:29:11.173935 4669 generic.go:334] "Generic (PLEG): container finished" podID="06094d4a-33e6-4b9f-b291-90262283b33a" containerID="3768ca9b47d1a830f88af92ef8cd83104e84dfc9021dd33f4521c7ce0a0f7460" exitCode=143 Oct 10 09:29:11 crc kubenswrapper[4669]: I1010 09:29:11.174135 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbc79ddd8-vqntz" event={"ID":"06094d4a-33e6-4b9f-b291-90262283b33a","Type":"ContainerDied","Data":"3768ca9b47d1a830f88af92ef8cd83104e84dfc9021dd33f4521c7ce0a0f7460"} Oct 10 09:29:12 crc kubenswrapper[4669]: I1010 09:29:12.245321 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:12 crc kubenswrapper[4669]: I1010 09:29:12.245579 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-central-agent" containerID="cri-o://6d961d1a9ef9537910a8a6e12d37020feadbdd40bba00d47c30ca513cc2e3e9e" gracePeriod=30 Oct 10 09:29:12 crc kubenswrapper[4669]: I1010 09:29:12.246237 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="sg-core" containerID="cri-o://2559692f8cbd2848abecd8a581a5fcee60d267ffaba17ae2aa3123d36e97cec3" gracePeriod=30 Oct 10 09:29:12 crc kubenswrapper[4669]: I1010 09:29:12.246283 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="proxy-httpd" containerID="cri-o://68a72109ef57ed0c00c70dcfea6e5b064bb1f0728abc24e418d154464647f1c8" gracePeriod=30 Oct 10 09:29:12 crc kubenswrapper[4669]: I1010 09:29:12.246297 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-notification-agent" containerID="cri-o://403adf74fc07ea0cf34dbcebe137f53a8ba5e31d91bf573e559672303cb82066" gracePeriod=30 Oct 10 09:29:12 crc kubenswrapper[4669]: I1010 09:29:12.262621 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.142:3000/\": EOF" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.191981 4669 generic.go:334] "Generic (PLEG): container finished" podID="a1d9870d-304c-4ab1-9091-9e7e939cc6b8" containerID="4e6e1e2d9fa7339fbcfd04e37c884a1eb4d3cbfc91ed126a8acf41e26a14a3dc" exitCode=0 Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.192194 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-trw89" event={"ID":"a1d9870d-304c-4ab1-9091-9e7e939cc6b8","Type":"ContainerDied","Data":"4e6e1e2d9fa7339fbcfd04e37c884a1eb4d3cbfc91ed126a8acf41e26a14a3dc"} Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.196072 4669 generic.go:334] "Generic (PLEG): container finished" podID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerID="68a72109ef57ed0c00c70dcfea6e5b064bb1f0728abc24e418d154464647f1c8" exitCode=0 Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.196094 4669 generic.go:334] "Generic (PLEG): container finished" podID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerID="2559692f8cbd2848abecd8a581a5fcee60d267ffaba17ae2aa3123d36e97cec3" exitCode=2 Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.196103 4669 generic.go:334] "Generic (PLEG): container finished" podID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerID="6d961d1a9ef9537910a8a6e12d37020feadbdd40bba00d47c30ca513cc2e3e9e" exitCode=0 Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.196110 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerDied","Data":"68a72109ef57ed0c00c70dcfea6e5b064bb1f0728abc24e418d154464647f1c8"} Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.196174 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerDied","Data":"2559692f8cbd2848abecd8a581a5fcee60d267ffaba17ae2aa3123d36e97cec3"} Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.196185 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerDied","Data":"6d961d1a9ef9537910a8a6e12d37020feadbdd40bba00d47c30ca513cc2e3e9e"} Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.477689 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-zwrh7"] Oct 10 09:29:13 crc kubenswrapper[4669]: E1010 09:29:13.478043 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f556bd-2a52-4976-8709-0d79392c793e" containerName="init" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.478054 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f556bd-2a52-4976-8709-0d79392c793e" containerName="init" Oct 10 09:29:13 crc kubenswrapper[4669]: E1010 09:29:13.478080 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f556bd-2a52-4976-8709-0d79392c793e" containerName="dnsmasq-dns" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.478087 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f556bd-2a52-4976-8709-0d79392c793e" containerName="dnsmasq-dns" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.478262 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="60f556bd-2a52-4976-8709-0d79392c793e" containerName="dnsmasq-dns" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.478809 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zwrh7" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.499891 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zwrh7"] Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.583719 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-l66br"] Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.585078 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l66br" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.604695 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-l66br"] Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.659735 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntvcq\" (UniqueName: \"kubernetes.io/projected/a4fe92a4-97c8-4787-94c3-89ae76827800-kube-api-access-ntvcq\") pod \"nova-api-db-create-zwrh7\" (UID: \"a4fe92a4-97c8-4787-94c3-89ae76827800\") " pod="openstack/nova-api-db-create-zwrh7" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.697677 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-qdk89"] Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.698767 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdk89" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.721655 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qdk89"] Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.761645 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntvcq\" (UniqueName: \"kubernetes.io/projected/a4fe92a4-97c8-4787-94c3-89ae76827800-kube-api-access-ntvcq\") pod \"nova-api-db-create-zwrh7\" (UID: \"a4fe92a4-97c8-4787-94c3-89ae76827800\") " pod="openstack/nova-api-db-create-zwrh7" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.761813 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbpjj\" (UniqueName: \"kubernetes.io/projected/343f638d-3bd8-4737-99ee-a94844f9a261-kube-api-access-wbpjj\") pod \"nova-cell0-db-create-l66br\" (UID: \"343f638d-3bd8-4737-99ee-a94844f9a261\") " pod="openstack/nova-cell0-db-create-l66br" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.782561 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntvcq\" (UniqueName: \"kubernetes.io/projected/a4fe92a4-97c8-4787-94c3-89ae76827800-kube-api-access-ntvcq\") pod \"nova-api-db-create-zwrh7\" (UID: \"a4fe92a4-97c8-4787-94c3-89ae76827800\") " pod="openstack/nova-api-db-create-zwrh7" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.805507 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zwrh7" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.867633 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbzlv\" (UniqueName: \"kubernetes.io/projected/d6725ca7-4c3f-4e08-b053-d8eeeef15394-kube-api-access-pbzlv\") pod \"nova-cell1-db-create-qdk89\" (UID: \"d6725ca7-4c3f-4e08-b053-d8eeeef15394\") " pod="openstack/nova-cell1-db-create-qdk89" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.867721 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbpjj\" (UniqueName: \"kubernetes.io/projected/343f638d-3bd8-4737-99ee-a94844f9a261-kube-api-access-wbpjj\") pod \"nova-cell0-db-create-l66br\" (UID: \"343f638d-3bd8-4737-99ee-a94844f9a261\") " pod="openstack/nova-cell0-db-create-l66br" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.885711 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbpjj\" (UniqueName: \"kubernetes.io/projected/343f638d-3bd8-4737-99ee-a94844f9a261-kube-api-access-wbpjj\") pod \"nova-cell0-db-create-l66br\" (UID: \"343f638d-3bd8-4737-99ee-a94844f9a261\") " pod="openstack/nova-cell0-db-create-l66br" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.933523 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l66br" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.969747 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbzlv\" (UniqueName: \"kubernetes.io/projected/d6725ca7-4c3f-4e08-b053-d8eeeef15394-kube-api-access-pbzlv\") pod \"nova-cell1-db-create-qdk89\" (UID: \"d6725ca7-4c3f-4e08-b053-d8eeeef15394\") " pod="openstack/nova-cell1-db-create-qdk89" Oct 10 09:29:13 crc kubenswrapper[4669]: I1010 09:29:13.991342 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbzlv\" (UniqueName: \"kubernetes.io/projected/d6725ca7-4c3f-4e08-b053-d8eeeef15394-kube-api-access-pbzlv\") pod \"nova-cell1-db-create-qdk89\" (UID: \"d6725ca7-4c3f-4e08-b053-d8eeeef15394\") " pod="openstack/nova-cell1-db-create-qdk89" Oct 10 09:29:14 crc kubenswrapper[4669]: I1010 09:29:14.023904 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdk89" Oct 10 09:29:14 crc kubenswrapper[4669]: I1010 09:29:14.090726 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cbc79ddd8-vqntz" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:59482->10.217.0.146:9311: read: connection reset by peer" Oct 10 09:29:14 crc kubenswrapper[4669]: I1010 09:29:14.091002 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cbc79ddd8-vqntz" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": read tcp 10.217.0.2:59484->10.217.0.146:9311: read: connection reset by peer" Oct 10 09:29:14 crc kubenswrapper[4669]: I1010 09:29:14.213022 4669 generic.go:334] "Generic (PLEG): container finished" podID="06094d4a-33e6-4b9f-b291-90262283b33a" containerID="c398199fc54e6e39ff3b3fa0e9e4e7d32fa49777018d2f46a5cbcb3c9cdd05d1" exitCode=0 Oct 10 09:29:14 crc kubenswrapper[4669]: I1010 09:29:14.213209 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbc79ddd8-vqntz" event={"ID":"06094d4a-33e6-4b9f-b291-90262283b33a","Type":"ContainerDied","Data":"c398199fc54e6e39ff3b3fa0e9e4e7d32fa49777018d2f46a5cbcb3c9cdd05d1"} Oct 10 09:29:16 crc kubenswrapper[4669]: I1010 09:29:16.252981 4669 generic.go:334] "Generic (PLEG): container finished" podID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerID="403adf74fc07ea0cf34dbcebe137f53a8ba5e31d91bf573e559672303cb82066" exitCode=0 Oct 10 09:29:16 crc kubenswrapper[4669]: I1010 09:29:16.253065 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerDied","Data":"403adf74fc07ea0cf34dbcebe137f53a8ba5e31d91bf573e559672303cb82066"} Oct 10 09:29:16 crc kubenswrapper[4669]: I1010 09:29:16.565853 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cbc79ddd8-vqntz" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": dial tcp 10.217.0.146:9311: connect: connection refused" Oct 10 09:29:16 crc kubenswrapper[4669]: I1010 09:29:16.565854 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cbc79ddd8-vqntz" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.146:9311/healthcheck\": dial tcp 10.217.0.146:9311: connect: connection refused" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.592505 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-trw89" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.653859 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-config-data\") pod \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.654205 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-scripts\") pod \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.654262 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-etc-machine-id\") pod \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.654323 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-db-sync-config-data\") pod \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.654419 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-combined-ca-bundle\") pod \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.654468 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t29ld\" (UniqueName: \"kubernetes.io/projected/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-kube-api-access-t29ld\") pod \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\" (UID: \"a1d9870d-304c-4ab1-9091-9e7e939cc6b8\") " Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.654874 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a1d9870d-304c-4ab1-9091-9e7e939cc6b8" (UID: "a1d9870d-304c-4ab1-9091-9e7e939cc6b8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.655278 4669 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.659181 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-scripts" (OuterVolumeSpecName: "scripts") pod "a1d9870d-304c-4ab1-9091-9e7e939cc6b8" (UID: "a1d9870d-304c-4ab1-9091-9e7e939cc6b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.661705 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a1d9870d-304c-4ab1-9091-9e7e939cc6b8" (UID: "a1d9870d-304c-4ab1-9091-9e7e939cc6b8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.664986 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-kube-api-access-t29ld" (OuterVolumeSpecName: "kube-api-access-t29ld") pod "a1d9870d-304c-4ab1-9091-9e7e939cc6b8" (UID: "a1d9870d-304c-4ab1-9091-9e7e939cc6b8"). InnerVolumeSpecName "kube-api-access-t29ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.763962 4669 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.763993 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t29ld\" (UniqueName: \"kubernetes.io/projected/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-kube-api-access-t29ld\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.764004 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.777610 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-config-data" (OuterVolumeSpecName: "config-data") pod "a1d9870d-304c-4ab1-9091-9e7e939cc6b8" (UID: "a1d9870d-304c-4ab1-9091-9e7e939cc6b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.856041 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1d9870d-304c-4ab1-9091-9e7e939cc6b8" (UID: "a1d9870d-304c-4ab1-9091-9e7e939cc6b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.869315 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.869356 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1d9870d-304c-4ab1-9091-9e7e939cc6b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:17 crc kubenswrapper[4669]: I1010 09:29:17.968815 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.002631 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074376 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-combined-ca-bundle\") pod \"06094d4a-33e6-4b9f-b291-90262283b33a\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074441 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-scripts\") pod \"fd93c04e-6db4-4411-8390-d10cd10b8048\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074475 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data-custom\") pod \"06094d4a-33e6-4b9f-b291-90262283b33a\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074500 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-run-httpd\") pod \"fd93c04e-6db4-4411-8390-d10cd10b8048\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074549 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-sg-core-conf-yaml\") pod \"fd93c04e-6db4-4411-8390-d10cd10b8048\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074616 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ml68\" (UniqueName: \"kubernetes.io/projected/fd93c04e-6db4-4411-8390-d10cd10b8048-kube-api-access-7ml68\") pod \"fd93c04e-6db4-4411-8390-d10cd10b8048\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074636 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-log-httpd\") pod \"fd93c04e-6db4-4411-8390-d10cd10b8048\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074674 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6dnj\" (UniqueName: \"kubernetes.io/projected/06094d4a-33e6-4b9f-b291-90262283b33a-kube-api-access-w6dnj\") pod \"06094d4a-33e6-4b9f-b291-90262283b33a\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074731 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data\") pod \"06094d4a-33e6-4b9f-b291-90262283b33a\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074791 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06094d4a-33e6-4b9f-b291-90262283b33a-logs\") pod \"06094d4a-33e6-4b9f-b291-90262283b33a\" (UID: \"06094d4a-33e6-4b9f-b291-90262283b33a\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074815 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-combined-ca-bundle\") pod \"fd93c04e-6db4-4411-8390-d10cd10b8048\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.074864 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-config-data\") pod \"fd93c04e-6db4-4411-8390-d10cd10b8048\" (UID: \"fd93c04e-6db4-4411-8390-d10cd10b8048\") " Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.081011 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd93c04e-6db4-4411-8390-d10cd10b8048-kube-api-access-7ml68" (OuterVolumeSpecName: "kube-api-access-7ml68") pod "fd93c04e-6db4-4411-8390-d10cd10b8048" (UID: "fd93c04e-6db4-4411-8390-d10cd10b8048"). InnerVolumeSpecName "kube-api-access-7ml68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.081311 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fd93c04e-6db4-4411-8390-d10cd10b8048" (UID: "fd93c04e-6db4-4411-8390-d10cd10b8048"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.086154 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "06094d4a-33e6-4b9f-b291-90262283b33a" (UID: "06094d4a-33e6-4b9f-b291-90262283b33a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.090648 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06094d4a-33e6-4b9f-b291-90262283b33a-logs" (OuterVolumeSpecName: "logs") pod "06094d4a-33e6-4b9f-b291-90262283b33a" (UID: "06094d4a-33e6-4b9f-b291-90262283b33a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.091843 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fd93c04e-6db4-4411-8390-d10cd10b8048" (UID: "fd93c04e-6db4-4411-8390-d10cd10b8048"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.111750 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-scripts" (OuterVolumeSpecName: "scripts") pod "fd93c04e-6db4-4411-8390-d10cd10b8048" (UID: "fd93c04e-6db4-4411-8390-d10cd10b8048"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.112036 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06094d4a-33e6-4b9f-b291-90262283b33a-kube-api-access-w6dnj" (OuterVolumeSpecName: "kube-api-access-w6dnj") pod "06094d4a-33e6-4b9f-b291-90262283b33a" (UID: "06094d4a-33e6-4b9f-b291-90262283b33a"). InnerVolumeSpecName "kube-api-access-w6dnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.163833 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fd93c04e-6db4-4411-8390-d10cd10b8048" (UID: "fd93c04e-6db4-4411-8390-d10cd10b8048"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.185924 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06094d4a-33e6-4b9f-b291-90262283b33a-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.185970 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.185981 4669 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.185991 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.186001 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.186013 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ml68\" (UniqueName: \"kubernetes.io/projected/fd93c04e-6db4-4411-8390-d10cd10b8048-kube-api-access-7ml68\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.186022 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd93c04e-6db4-4411-8390-d10cd10b8048-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.186030 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6dnj\" (UniqueName: \"kubernetes.io/projected/06094d4a-33e6-4b9f-b291-90262283b33a-kube-api-access-w6dnj\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.186722 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06094d4a-33e6-4b9f-b291-90262283b33a" (UID: "06094d4a-33e6-4b9f-b291-90262283b33a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.222120 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd93c04e-6db4-4411-8390-d10cd10b8048" (UID: "fd93c04e-6db4-4411-8390-d10cd10b8048"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.227841 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zwrh7"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.269073 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data" (OuterVolumeSpecName: "config-data") pod "06094d4a-33e6-4b9f-b291-90262283b33a" (UID: "06094d4a-33e6-4b9f-b291-90262283b33a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.289578 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.290673 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06094d4a-33e6-4b9f-b291-90262283b33a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.290687 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.293863 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd93c04e-6db4-4411-8390-d10cd10b8048","Type":"ContainerDied","Data":"a0a7f9c7a2d8fc270afbab41fccdb7505f661651003ccaf3fc5b0e250d008b74"} Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.293908 4669 scope.go:117] "RemoveContainer" containerID="68a72109ef57ed0c00c70dcfea6e5b064bb1f0728abc24e418d154464647f1c8" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.294094 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.297246 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2bb8c92a-28c8-4424-94c2-c577f0b6b486","Type":"ContainerStarted","Data":"f6f2d7a6cbeab2d9b86fad60521f34476df800d18831d4d02ff52718760b5a8c"} Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.314212 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbc79ddd8-vqntz" event={"ID":"06094d4a-33e6-4b9f-b291-90262283b33a","Type":"ContainerDied","Data":"84666ba9a358ae2048259324909f066d5fcd89d8776f8cf9c597c315a6331293"} Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.314331 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cbc79ddd8-vqntz" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.322751 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zwrh7" event={"ID":"a4fe92a4-97c8-4787-94c3-89ae76827800","Type":"ContainerStarted","Data":"1f9ba32c5aba99d407a9637a9f7ac01f8a9ae7cff53dca715ff145a424f1fe7e"} Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.339170 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.096739508 podStartE2EDuration="16.339151167s" podCreationTimestamp="2025-10-10 09:29:02 +0000 UTC" firstStartedPulling="2025-10-10 09:29:03.301113472 +0000 UTC m=+1086.317132214" lastFinishedPulling="2025-10-10 09:29:17.543525141 +0000 UTC m=+1100.559543873" observedRunningTime="2025-10-10 09:29:18.326250842 +0000 UTC m=+1101.342269594" watchObservedRunningTime="2025-10-10 09:29:18.339151167 +0000 UTC m=+1101.355169899" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.341040 4669 scope.go:117] "RemoveContainer" containerID="2559692f8cbd2848abecd8a581a5fcee60d267ffaba17ae2aa3123d36e97cec3" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.351800 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-trw89" event={"ID":"a1d9870d-304c-4ab1-9091-9e7e939cc6b8","Type":"ContainerDied","Data":"4c1ae6a70585cb9226fcfdac265a0e882b17cca33ee557c6a7b3baeee9fc2b56"} Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.351838 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c1ae6a70585cb9226fcfdac265a0e882b17cca33ee557c6a7b3baeee9fc2b56" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.351913 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-trw89" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.398275 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-config-data" (OuterVolumeSpecName: "config-data") pod "fd93c04e-6db4-4411-8390-d10cd10b8048" (UID: "fd93c04e-6db4-4411-8390-d10cd10b8048"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.433854 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qdk89"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.469389 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-l66br"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.500694 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd93c04e-6db4-4411-8390-d10cd10b8048-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.516651 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-cbc79ddd8-vqntz"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.524903 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-cbc79ddd8-vqntz"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.530792 4669 scope.go:117] "RemoveContainer" containerID="403adf74fc07ea0cf34dbcebe137f53a8ba5e31d91bf573e559672303cb82066" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.622861 4669 scope.go:117] "RemoveContainer" containerID="6d961d1a9ef9537910a8a6e12d37020feadbdd40bba00d47c30ca513cc2e3e9e" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.651782 4669 scope.go:117] "RemoveContainer" containerID="c398199fc54e6e39ff3b3fa0e9e4e7d32fa49777018d2f46a5cbcb3c9cdd05d1" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.662789 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.679669 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.687556 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:18 crc kubenswrapper[4669]: E1010 09:29:18.689152 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-central-agent" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689172 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-central-agent" Oct 10 09:29:18 crc kubenswrapper[4669]: E1010 09:29:18.689192 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-notification-agent" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689198 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-notification-agent" Oct 10 09:29:18 crc kubenswrapper[4669]: E1010 09:29:18.689210 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api-log" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689215 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api-log" Oct 10 09:29:18 crc kubenswrapper[4669]: E1010 09:29:18.689233 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="proxy-httpd" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689239 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="proxy-httpd" Oct 10 09:29:18 crc kubenswrapper[4669]: E1010 09:29:18.689248 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="sg-core" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689253 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="sg-core" Oct 10 09:29:18 crc kubenswrapper[4669]: E1010 09:29:18.689263 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689268 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api" Oct 10 09:29:18 crc kubenswrapper[4669]: E1010 09:29:18.689278 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1d9870d-304c-4ab1-9091-9e7e939cc6b8" containerName="cinder-db-sync" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689283 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1d9870d-304c-4ab1-9091-9e7e939cc6b8" containerName="cinder-db-sync" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689432 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1d9870d-304c-4ab1-9091-9e7e939cc6b8" containerName="cinder-db-sync" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689446 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="proxy-httpd" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689454 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-central-agent" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689463 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="ceilometer-notification-agent" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689473 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" containerName="sg-core" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689482 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api-log" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.689495 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" containerName="barbican-api" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.691309 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.691634 4669 scope.go:117] "RemoveContainer" containerID="3768ca9b47d1a830f88af92ef8cd83104e84dfc9021dd33f4521c7ce0a0f7460" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.695946 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.696088 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.705983 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.805200 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-config-data\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.805425 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-log-httpd\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.805548 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2kgf\" (UniqueName: \"kubernetes.io/projected/16462469-a4fc-4197-af0e-f9d4973b6564-kube-api-access-w2kgf\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.805684 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.805769 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.805834 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-run-httpd\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.806092 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-scripts\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.885233 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.886988 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.891249 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.891415 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.891790 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jpczk" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.891907 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.907736 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.907787 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.907839 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.907887 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-scripts\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.907933 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8wgn\" (UniqueName: \"kubernetes.io/projected/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-kube-api-access-c8wgn\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.907987 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-config-data\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.908001 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.908029 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.908047 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-log-httpd\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.908064 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2kgf\" (UniqueName: \"kubernetes.io/projected/16462469-a4fc-4197-af0e-f9d4973b6564-kube-api-access-w2kgf\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.908085 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.908104 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.908120 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-run-httpd\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.910701 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-log-httpd\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.911375 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-run-httpd\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.915573 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-config-data\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.915875 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.916246 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-scripts\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.917277 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.937123 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.964197 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2kgf\" (UniqueName: \"kubernetes.io/projected/16462469-a4fc-4197-af0e-f9d4973b6564-kube-api-access-w2kgf\") pod \"ceilometer-0\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " pod="openstack/ceilometer-0" Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.988090 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85494b87f-df9tj"] Oct 10 09:29:18 crc kubenswrapper[4669]: I1010 09:29:18.989321 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.009476 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.009515 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-nb\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.009552 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.009573 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cclz8\" (UniqueName: \"kubernetes.io/projected/6eb84575-ef26-4a1e-abe1-326f39ecaddf-kube-api-access-cclz8\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.009629 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-sb\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.009667 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8wgn\" (UniqueName: \"kubernetes.io/projected/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-kube-api-access-c8wgn\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.009702 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-dns-svc\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.010163 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.010198 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.010220 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-config\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.010251 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.014241 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.020745 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.027074 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.031801 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-df9tj"] Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.033275 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.037610 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.049192 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8wgn\" (UniqueName: \"kubernetes.io/projected/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-kube-api-access-c8wgn\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.052043 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.118723 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cclz8\" (UniqueName: \"kubernetes.io/projected/6eb84575-ef26-4a1e-abe1-326f39ecaddf-kube-api-access-cclz8\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.118780 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-sb\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.118823 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-dns-svc\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.118856 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-config\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.118894 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-nb\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.119662 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-nb\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.120368 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-sb\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.123571 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-config\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.142055 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-dns-svc\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.174289 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cclz8\" (UniqueName: \"kubernetes.io/projected/6eb84575-ef26-4a1e-abe1-326f39ecaddf-kube-api-access-cclz8\") pod \"dnsmasq-dns-85494b87f-df9tj\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.207022 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.274677 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.276464 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.286270 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.316814 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.318077 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.330647 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data-custom\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.330872 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.330964 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4ad01b-73e6-4a76-a788-952772dbf376-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.331085 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fw48\" (UniqueName: \"kubernetes.io/projected/5b4ad01b-73e6-4a76-a788-952772dbf376-kube-api-access-7fw48\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.331153 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-scripts\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.331260 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b4ad01b-73e6-4a76-a788-952772dbf376-logs\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.331324 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.438399 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data-custom\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.438467 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.438534 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4ad01b-73e6-4a76-a788-952772dbf376-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.438611 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fw48\" (UniqueName: \"kubernetes.io/projected/5b4ad01b-73e6-4a76-a788-952772dbf376-kube-api-access-7fw48\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.438655 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-scripts\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.438713 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b4ad01b-73e6-4a76-a788-952772dbf376-logs\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.438751 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.439363 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4ad01b-73e6-4a76-a788-952772dbf376-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.439951 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b4ad01b-73e6-4a76-a788-952772dbf376-logs\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.453792 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-scripts\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.459358 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.460025 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.463194 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data-custom\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.470776 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fw48\" (UniqueName: \"kubernetes.io/projected/5b4ad01b-73e6-4a76-a788-952772dbf376-kube-api-access-7fw48\") pod \"cinder-api-0\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.549615 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdk89" event={"ID":"d6725ca7-4c3f-4e08-b053-d8eeeef15394","Type":"ContainerStarted","Data":"763bc4a1878d058c147ff64408d44018e9014c82686cac8366fafd168f150979"} Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.549657 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdk89" event={"ID":"d6725ca7-4c3f-4e08-b053-d8eeeef15394","Type":"ContainerStarted","Data":"1452f5e33253666dfb695207d0b09216d866f533fa0b78216df1d5ab05968e20"} Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.555120 4669 generic.go:334] "Generic (PLEG): container finished" podID="a4fe92a4-97c8-4787-94c3-89ae76827800" containerID="0f0a4e06557b0dbdd9cc9514819fe08cc24eff230edfea029fd9994358e39be6" exitCode=0 Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.555307 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zwrh7" event={"ID":"a4fe92a4-97c8-4787-94c3-89ae76827800","Type":"ContainerDied","Data":"0f0a4e06557b0dbdd9cc9514819fe08cc24eff230edfea029fd9994358e39be6"} Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.557422 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l66br" event={"ID":"343f638d-3bd8-4737-99ee-a94844f9a261","Type":"ContainerStarted","Data":"218bc52fe18fe937533de4bc76efd2d255c6b498fa98167ebc7c538bf8d8bb3b"} Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.566467 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l66br" event={"ID":"343f638d-3bd8-4737-99ee-a94844f9a261","Type":"ContainerStarted","Data":"249300d3dd8be00e454ddb0ce24c120eb6d19194df73b149dec7943c8df14505"} Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.606405 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-qdk89" podStartSLOduration=6.606367454 podStartE2EDuration="6.606367454s" podCreationTimestamp="2025-10-10 09:29:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:19.59194184 +0000 UTC m=+1102.607960582" watchObservedRunningTime="2025-10-10 09:29:19.606367454 +0000 UTC m=+1102.622386196" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.710808 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-l66br" podStartSLOduration=6.710786181 podStartE2EDuration="6.710786181s" podCreationTimestamp="2025-10-10 09:29:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:19.676442817 +0000 UTC m=+1102.692461559" watchObservedRunningTime="2025-10-10 09:29:19.710786181 +0000 UTC m=+1102.726804913" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.718456 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.832718 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06094d4a-33e6-4b9f-b291-90262283b33a" path="/var/lib/kubelet/pods/06094d4a-33e6-4b9f-b291-90262283b33a/volumes" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.833839 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd93c04e-6db4-4411-8390-d10cd10b8048" path="/var/lib/kubelet/pods/fd93c04e-6db4-4411-8390-d10cd10b8048/volumes" Oct 10 09:29:19 crc kubenswrapper[4669]: I1010 09:29:19.981868 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.254966 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-df9tj"] Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.266209 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:20 crc kubenswrapper[4669]: W1010 09:29:20.267303 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae8ada7a_c97b_423d_b249_bc7c98d8b8b6.slice/crio-5d3ca8332495dfc2fb09e01a8a8712e6c52db25fc6009c084efb41a649484a33 WatchSource:0}: Error finding container 5d3ca8332495dfc2fb09e01a8a8712e6c52db25fc6009c084efb41a649484a33: Status 404 returned error can't find the container with id 5d3ca8332495dfc2fb09e01a8a8712e6c52db25fc6009c084efb41a649484a33 Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.384964 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:20 crc kubenswrapper[4669]: W1010 09:29:20.393763 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b4ad01b_73e6_4a76_a788_952772dbf376.slice/crio-28ef63bc20b18ed81c620c32dd762c2f9c81d2f3dec51c89906b0fee9d267c39 WatchSource:0}: Error finding container 28ef63bc20b18ed81c620c32dd762c2f9c81d2f3dec51c89906b0fee9d267c39: Status 404 returned error can't find the container with id 28ef63bc20b18ed81c620c32dd762c2f9c81d2f3dec51c89906b0fee9d267c39 Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.585042 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-df9tj" event={"ID":"6eb84575-ef26-4a1e-abe1-326f39ecaddf","Type":"ContainerStarted","Data":"a73bb82b38507e41ff46d643ace36481a49c312934299a7053d9346985c19227"} Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.601597 4669 generic.go:334] "Generic (PLEG): container finished" podID="d6725ca7-4c3f-4e08-b053-d8eeeef15394" containerID="763bc4a1878d058c147ff64408d44018e9014c82686cac8366fafd168f150979" exitCode=0 Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.601712 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdk89" event={"ID":"d6725ca7-4c3f-4e08-b053-d8eeeef15394","Type":"ContainerDied","Data":"763bc4a1878d058c147ff64408d44018e9014c82686cac8366fafd168f150979"} Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.604714 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5b4ad01b-73e6-4a76-a788-952772dbf376","Type":"ContainerStarted","Data":"28ef63bc20b18ed81c620c32dd762c2f9c81d2f3dec51c89906b0fee9d267c39"} Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.606042 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6","Type":"ContainerStarted","Data":"5d3ca8332495dfc2fb09e01a8a8712e6c52db25fc6009c084efb41a649484a33"} Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.618998 4669 generic.go:334] "Generic (PLEG): container finished" podID="343f638d-3bd8-4737-99ee-a94844f9a261" containerID="218bc52fe18fe937533de4bc76efd2d255c6b498fa98167ebc7c538bf8d8bb3b" exitCode=0 Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.619077 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l66br" event={"ID":"343f638d-3bd8-4737-99ee-a94844f9a261","Type":"ContainerDied","Data":"218bc52fe18fe937533de4bc76efd2d255c6b498fa98167ebc7c538bf8d8bb3b"} Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.630264 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerStarted","Data":"aaca94c8915ffb5160aa979d95ba136bb3a58454c3e645d73602006498724bf3"} Oct 10 09:29:20 crc kubenswrapper[4669]: I1010 09:29:20.960415 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zwrh7" Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.023137 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntvcq\" (UniqueName: \"kubernetes.io/projected/a4fe92a4-97c8-4787-94c3-89ae76827800-kube-api-access-ntvcq\") pod \"a4fe92a4-97c8-4787-94c3-89ae76827800\" (UID: \"a4fe92a4-97c8-4787-94c3-89ae76827800\") " Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.029195 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4fe92a4-97c8-4787-94c3-89ae76827800-kube-api-access-ntvcq" (OuterVolumeSpecName: "kube-api-access-ntvcq") pod "a4fe92a4-97c8-4787-94c3-89ae76827800" (UID: "a4fe92a4-97c8-4787-94c3-89ae76827800"). InnerVolumeSpecName "kube-api-access-ntvcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.125212 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntvcq\" (UniqueName: \"kubernetes.io/projected/a4fe92a4-97c8-4787-94c3-89ae76827800-kube-api-access-ntvcq\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.649525 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerStarted","Data":"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2"} Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.651327 4669 generic.go:334] "Generic (PLEG): container finished" podID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerID="df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5" exitCode=0 Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.651367 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-df9tj" event={"ID":"6eb84575-ef26-4a1e-abe1-326f39ecaddf","Type":"ContainerDied","Data":"df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5"} Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.655536 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zwrh7" event={"ID":"a4fe92a4-97c8-4787-94c3-89ae76827800","Type":"ContainerDied","Data":"1f9ba32c5aba99d407a9637a9f7ac01f8a9ae7cff53dca715ff145a424f1fe7e"} Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.655568 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f9ba32c5aba99d407a9637a9f7ac01f8a9ae7cff53dca715ff145a424f1fe7e" Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.655632 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zwrh7" Oct 10 09:29:21 crc kubenswrapper[4669]: I1010 09:29:21.659568 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5b4ad01b-73e6-4a76-a788-952772dbf376","Type":"ContainerStarted","Data":"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab"} Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.026293 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.273658 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdk89" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.297526 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l66br" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.360049 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbzlv\" (UniqueName: \"kubernetes.io/projected/d6725ca7-4c3f-4e08-b053-d8eeeef15394-kube-api-access-pbzlv\") pod \"d6725ca7-4c3f-4e08-b053-d8eeeef15394\" (UID: \"d6725ca7-4c3f-4e08-b053-d8eeeef15394\") " Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.360195 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbpjj\" (UniqueName: \"kubernetes.io/projected/343f638d-3bd8-4737-99ee-a94844f9a261-kube-api-access-wbpjj\") pod \"343f638d-3bd8-4737-99ee-a94844f9a261\" (UID: \"343f638d-3bd8-4737-99ee-a94844f9a261\") " Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.364652 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6725ca7-4c3f-4e08-b053-d8eeeef15394-kube-api-access-pbzlv" (OuterVolumeSpecName: "kube-api-access-pbzlv") pod "d6725ca7-4c3f-4e08-b053-d8eeeef15394" (UID: "d6725ca7-4c3f-4e08-b053-d8eeeef15394"). InnerVolumeSpecName "kube-api-access-pbzlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.377181 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/343f638d-3bd8-4737-99ee-a94844f9a261-kube-api-access-wbpjj" (OuterVolumeSpecName: "kube-api-access-wbpjj") pod "343f638d-3bd8-4737-99ee-a94844f9a261" (UID: "343f638d-3bd8-4737-99ee-a94844f9a261"). InnerVolumeSpecName "kube-api-access-wbpjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.462784 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbzlv\" (UniqueName: \"kubernetes.io/projected/d6725ca7-4c3f-4e08-b053-d8eeeef15394-kube-api-access-pbzlv\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.462813 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbpjj\" (UniqueName: \"kubernetes.io/projected/343f638d-3bd8-4737-99ee-a94844f9a261-kube-api-access-wbpjj\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.492862 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.766304 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-l66br" event={"ID":"343f638d-3bd8-4737-99ee-a94844f9a261","Type":"ContainerDied","Data":"249300d3dd8be00e454ddb0ce24c120eb6d19194df73b149dec7943c8df14505"} Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.766665 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="249300d3dd8be00e454ddb0ce24c120eb6d19194df73b149dec7943c8df14505" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.766349 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-l66br" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.809782 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerStarted","Data":"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46"} Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.836123 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.852126 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qdk89" event={"ID":"d6725ca7-4c3f-4e08-b053-d8eeeef15394","Type":"ContainerDied","Data":"1452f5e33253666dfb695207d0b09216d866f533fa0b78216df1d5ab05968e20"} Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.852163 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1452f5e33253666dfb695207d0b09216d866f533fa0b78216df1d5ab05968e20" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.852224 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qdk89" Oct 10 09:29:22 crc kubenswrapper[4669]: I1010 09:29:22.876063 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85494b87f-df9tj" podStartSLOduration=4.876040271 podStartE2EDuration="4.876040271s" podCreationTimestamp="2025-10-10 09:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:22.86698145 +0000 UTC m=+1105.883000192" watchObservedRunningTime="2025-10-10 09:29:22.876040271 +0000 UTC m=+1105.892059013" Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.861593 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5b4ad01b-73e6-4a76-a788-952772dbf376","Type":"ContainerStarted","Data":"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d"} Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.861626 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api-log" containerID="cri-o://8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab" gracePeriod=30 Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.862024 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.861713 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api" containerID="cri-o://ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d" gracePeriod=30 Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.868001 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6","Type":"ContainerStarted","Data":"811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba"} Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.868250 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6","Type":"ContainerStarted","Data":"7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305"} Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.870833 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerStarted","Data":"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7"} Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.872886 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-df9tj" event={"ID":"6eb84575-ef26-4a1e-abe1-326f39ecaddf","Type":"ContainerStarted","Data":"a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c"} Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.901620 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.9016029759999995 podStartE2EDuration="4.901602976s" podCreationTimestamp="2025-10-10 09:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:23.898528507 +0000 UTC m=+1106.914547249" watchObservedRunningTime="2025-10-10 09:29:23.901602976 +0000 UTC m=+1106.917621728" Oct 10 09:29:23 crc kubenswrapper[4669]: I1010 09:29:23.930025 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.804009115 podStartE2EDuration="5.930009439s" podCreationTimestamp="2025-10-10 09:29:18 +0000 UTC" firstStartedPulling="2025-10-10 09:29:20.271727466 +0000 UTC m=+1103.287746208" lastFinishedPulling="2025-10-10 09:29:21.39772779 +0000 UTC m=+1104.413746532" observedRunningTime="2025-10-10 09:29:23.925443582 +0000 UTC m=+1106.941462324" watchObservedRunningTime="2025-10-10 09:29:23.930009439 +0000 UTC m=+1106.946028181" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.210827 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.274873 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.275140 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.713191 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818073 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-combined-ca-bundle\") pod \"5b4ad01b-73e6-4a76-a788-952772dbf376\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818280 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data-custom\") pod \"5b4ad01b-73e6-4a76-a788-952772dbf376\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818322 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4ad01b-73e6-4a76-a788-952772dbf376-etc-machine-id\") pod \"5b4ad01b-73e6-4a76-a788-952772dbf376\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818354 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data\") pod \"5b4ad01b-73e6-4a76-a788-952772dbf376\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818373 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b4ad01b-73e6-4a76-a788-952772dbf376-logs\") pod \"5b4ad01b-73e6-4a76-a788-952772dbf376\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818411 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fw48\" (UniqueName: \"kubernetes.io/projected/5b4ad01b-73e6-4a76-a788-952772dbf376-kube-api-access-7fw48\") pod \"5b4ad01b-73e6-4a76-a788-952772dbf376\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818807 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b4ad01b-73e6-4a76-a788-952772dbf376-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5b4ad01b-73e6-4a76-a788-952772dbf376" (UID: "5b4ad01b-73e6-4a76-a788-952772dbf376"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.818853 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b4ad01b-73e6-4a76-a788-952772dbf376-logs" (OuterVolumeSpecName: "logs") pod "5b4ad01b-73e6-4a76-a788-952772dbf376" (UID: "5b4ad01b-73e6-4a76-a788-952772dbf376"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.819046 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-scripts\") pod \"5b4ad01b-73e6-4a76-a788-952772dbf376\" (UID: \"5b4ad01b-73e6-4a76-a788-952772dbf376\") " Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.819764 4669 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5b4ad01b-73e6-4a76-a788-952772dbf376-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.819787 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b4ad01b-73e6-4a76-a788-952772dbf376-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.823768 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5b4ad01b-73e6-4a76-a788-952772dbf376" (UID: "5b4ad01b-73e6-4a76-a788-952772dbf376"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.828443 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b4ad01b-73e6-4a76-a788-952772dbf376-kube-api-access-7fw48" (OuterVolumeSpecName: "kube-api-access-7fw48") pod "5b4ad01b-73e6-4a76-a788-952772dbf376" (UID: "5b4ad01b-73e6-4a76-a788-952772dbf376"). InnerVolumeSpecName "kube-api-access-7fw48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.845763 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-scripts" (OuterVolumeSpecName: "scripts") pod "5b4ad01b-73e6-4a76-a788-952772dbf376" (UID: "5b4ad01b-73e6-4a76-a788-952772dbf376"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.855323 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b4ad01b-73e6-4a76-a788-952772dbf376" (UID: "5b4ad01b-73e6-4a76-a788-952772dbf376"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.877341 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data" (OuterVolumeSpecName: "config-data") pod "5b4ad01b-73e6-4a76-a788-952772dbf376" (UID: "5b4ad01b-73e6-4a76-a788-952772dbf376"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.882948 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerStarted","Data":"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71"} Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.883693 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.888654 4669 generic.go:334] "Generic (PLEG): container finished" podID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerID="ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d" exitCode=0 Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.888677 4669 generic.go:334] "Generic (PLEG): container finished" podID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerID="8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab" exitCode=143 Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.889049 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5b4ad01b-73e6-4a76-a788-952772dbf376","Type":"ContainerDied","Data":"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d"} Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.889843 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.889943 4669 scope.go:117] "RemoveContainer" containerID="ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.896235 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5b4ad01b-73e6-4a76-a788-952772dbf376","Type":"ContainerDied","Data":"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab"} Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.896294 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5b4ad01b-73e6-4a76-a788-952772dbf376","Type":"ContainerDied","Data":"28ef63bc20b18ed81c620c32dd762c2f9c81d2f3dec51c89906b0fee9d267c39"} Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.913725 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9933781379999997 podStartE2EDuration="6.913707617s" podCreationTimestamp="2025-10-10 09:29:18 +0000 UTC" firstStartedPulling="2025-10-10 09:29:19.991417803 +0000 UTC m=+1103.007436545" lastFinishedPulling="2025-10-10 09:29:23.911747282 +0000 UTC m=+1106.927766024" observedRunningTime="2025-10-10 09:29:24.905293676 +0000 UTC m=+1107.921312418" watchObservedRunningTime="2025-10-10 09:29:24.913707617 +0000 UTC m=+1107.929726349" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.921342 4669 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.921374 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.921385 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fw48\" (UniqueName: \"kubernetes.io/projected/5b4ad01b-73e6-4a76-a788-952772dbf376-kube-api-access-7fw48\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.921394 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.921403 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4ad01b-73e6-4a76-a788-952772dbf376-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.955934 4669 scope.go:117] "RemoveContainer" containerID="8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.973521 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.987451 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.994221 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:24 crc kubenswrapper[4669]: E1010 09:29:24.994684 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.994713 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api" Oct 10 09:29:24 crc kubenswrapper[4669]: E1010 09:29:24.994740 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6725ca7-4c3f-4e08-b053-d8eeeef15394" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.994749 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6725ca7-4c3f-4e08-b053-d8eeeef15394" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: E1010 09:29:24.994785 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343f638d-3bd8-4737-99ee-a94844f9a261" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.994795 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="343f638d-3bd8-4737-99ee-a94844f9a261" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: E1010 09:29:24.994809 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4fe92a4-97c8-4787-94c3-89ae76827800" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.994817 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4fe92a4-97c8-4787-94c3-89ae76827800" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: E1010 09:29:24.994827 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api-log" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.994834 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api-log" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.995002 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4fe92a4-97c8-4787-94c3-89ae76827800" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.995021 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api-log" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.995032 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="343f638d-3bd8-4737-99ee-a94844f9a261" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.995040 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" containerName="cinder-api" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.995061 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6725ca7-4c3f-4e08-b053-d8eeeef15394" containerName="mariadb-database-create" Oct 10 09:29:24 crc kubenswrapper[4669]: I1010 09:29:24.996060 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.012497 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.012856 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.013154 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.024050 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.024664 4669 scope.go:117] "RemoveContainer" containerID="ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d" Oct 10 09:29:25 crc kubenswrapper[4669]: E1010 09:29:25.028758 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d\": container with ID starting with ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d not found: ID does not exist" containerID="ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.028799 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d"} err="failed to get container status \"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d\": rpc error: code = NotFound desc = could not find container \"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d\": container with ID starting with ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d not found: ID does not exist" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.028827 4669 scope.go:117] "RemoveContainer" containerID="8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab" Oct 10 09:29:25 crc kubenswrapper[4669]: E1010 09:29:25.031866 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab\": container with ID starting with 8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab not found: ID does not exist" containerID="8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.031909 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab"} err="failed to get container status \"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab\": rpc error: code = NotFound desc = could not find container \"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab\": container with ID starting with 8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab not found: ID does not exist" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.031938 4669 scope.go:117] "RemoveContainer" containerID="ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.032409 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d"} err="failed to get container status \"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d\": rpc error: code = NotFound desc = could not find container \"ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d\": container with ID starting with ea065649918303153a29d80fbc1636999ecda1e5cd66930ea7f0bf5aab8bab5d not found: ID does not exist" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.032431 4669 scope.go:117] "RemoveContainer" containerID="8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.032687 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab"} err="failed to get container status \"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab\": rpc error: code = NotFound desc = could not find container \"8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab\": container with ID starting with 8beb89a1effa91048051e091d089639aef9f06f3878851d25840f739012ec0ab not found: ID does not exist" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127252 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-scripts\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127385 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-config-data\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127418 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127466 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gdjd\" (UniqueName: \"kubernetes.io/projected/50ffa58f-0d75-4996-b66c-092aacc84c63-kube-api-access-9gdjd\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127485 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-config-data-custom\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127514 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ffa58f-0d75-4996-b66c-092aacc84c63-logs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127560 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127601 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ffa58f-0d75-4996-b66c-092aacc84c63-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.127619 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-public-tls-certs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229330 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-config-data-custom\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229385 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ffa58f-0d75-4996-b66c-092aacc84c63-logs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229454 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229485 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ffa58f-0d75-4996-b66c-092aacc84c63-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229514 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-public-tls-certs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229532 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-scripts\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229567 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-config-data\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229600 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.229642 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gdjd\" (UniqueName: \"kubernetes.io/projected/50ffa58f-0d75-4996-b66c-092aacc84c63-kube-api-access-9gdjd\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.230761 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50ffa58f-0d75-4996-b66c-092aacc84c63-logs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.231063 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/50ffa58f-0d75-4996-b66c-092aacc84c63-etc-machine-id\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.234147 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-config-data-custom\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.235806 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-scripts\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.235925 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-public-tls-certs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.236250 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.236556 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.243755 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50ffa58f-0d75-4996-b66c-092aacc84c63-config-data\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.251450 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gdjd\" (UniqueName: \"kubernetes.io/projected/50ffa58f-0d75-4996-b66c-092aacc84c63-kube-api-access-9gdjd\") pod \"cinder-api-0\" (UID: \"50ffa58f-0d75-4996-b66c-092aacc84c63\") " pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.347441 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.805487 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b4ad01b-73e6-4a76-a788-952772dbf376" path="/var/lib/kubelet/pods/5b4ad01b-73e6-4a76-a788-952772dbf376/volumes" Oct 10 09:29:25 crc kubenswrapper[4669]: I1010 09:29:25.902054 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.279249 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6d556ffc9f-2x4qt" Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.371789 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-588b8ffdfb-2xrm4"] Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.372322 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-588b8ffdfb-2xrm4" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-api" containerID="cri-o://1ea8850ed5c4772058ca1f9fe9b9d04a3cb3fb182cd1cfd9bd6c6673f5f00026" gracePeriod=30 Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.372819 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-588b8ffdfb-2xrm4" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-httpd" containerID="cri-o://272a322b633e4955e9bedb2be9f7cbb125c4b0428dd4109d2d1ab55e79bdc15a" gracePeriod=30 Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.950113 4669 generic.go:334] "Generic (PLEG): container finished" podID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerID="272a322b633e4955e9bedb2be9f7cbb125c4b0428dd4109d2d1ab55e79bdc15a" exitCode=0 Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.950249 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-588b8ffdfb-2xrm4" event={"ID":"c92807a3-b0ae-48c0-ba7b-1e341aabde2a","Type":"ContainerDied","Data":"272a322b633e4955e9bedb2be9f7cbb125c4b0428dd4109d2d1ab55e79bdc15a"} Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.956793 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ffa58f-0d75-4996-b66c-092aacc84c63","Type":"ContainerStarted","Data":"c93a339348f2183bb32f41b839fa9607c543e694666903e35a832d482583d3e3"} Oct 10 09:29:26 crc kubenswrapper[4669]: I1010 09:29:26.956851 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ffa58f-0d75-4996-b66c-092aacc84c63","Type":"ContainerStarted","Data":"9c1dcc1fa7582dba95742e12dd06e8411343a22b80bd99752d1318880f210edb"} Oct 10 09:29:27 crc kubenswrapper[4669]: I1010 09:29:27.966257 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"50ffa58f-0d75-4996-b66c-092aacc84c63","Type":"ContainerStarted","Data":"5d165405ec327a2893387958472e8bd8f5b6d1f2b9b59af964f6e494b88b7ce6"} Oct 10 09:29:27 crc kubenswrapper[4669]: I1010 09:29:27.966621 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 09:29:27 crc kubenswrapper[4669]: I1010 09:29:27.987894 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.987877159 podStartE2EDuration="3.987877159s" podCreationTimestamp="2025-10-10 09:29:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:27.986694511 +0000 UTC m=+1111.002713253" watchObservedRunningTime="2025-10-10 09:29:27.987877159 +0000 UTC m=+1111.003895901" Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.238136 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.238450 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-central-agent" containerID="cri-o://70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2" gracePeriod=30 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.238619 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="proxy-httpd" containerID="cri-o://c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71" gracePeriod=30 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.238665 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="sg-core" containerID="cri-o://b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7" gracePeriod=30 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.238706 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-notification-agent" containerID="cri-o://032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46" gracePeriod=30 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.320782 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.418579 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mgpjg"] Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.418840 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" podUID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerName="dnsmasq-dns" containerID="cri-o://ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec" gracePeriod=10 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.621415 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.702345 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.970659 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.980464 4669 generic.go:334] "Generic (PLEG): container finished" podID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerID="ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec" exitCode=0 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.980517 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" event={"ID":"eb84008d-41fb-46a0-97d9-4237d4575fe6","Type":"ContainerDied","Data":"ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec"} Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.980541 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" event={"ID":"eb84008d-41fb-46a0-97d9-4237d4575fe6","Type":"ContainerDied","Data":"bdcc73d3b4c24190017019770ddb3d2a2174085d33bb17c53969a736245b1c74"} Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.980558 4669 scope.go:117] "RemoveContainer" containerID="ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec" Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.980664 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mgpjg" Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.986147 4669 generic.go:334] "Generic (PLEG): container finished" podID="16462469-a4fc-4197-af0e-f9d4973b6564" containerID="c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71" exitCode=0 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.986192 4669 generic.go:334] "Generic (PLEG): container finished" podID="16462469-a4fc-4197-af0e-f9d4973b6564" containerID="b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7" exitCode=2 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.986201 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerDied","Data":"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71"} Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.986244 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerDied","Data":"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7"} Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.986344 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="cinder-scheduler" containerID="cri-o://7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305" gracePeriod=30 Oct 10 09:29:29 crc kubenswrapper[4669]: I1010 09:29:29.986424 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="probe" containerID="cri-o://811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba" gracePeriod=30 Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.017637 4669 scope.go:117] "RemoveContainer" containerID="0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.052387 4669 scope.go:117] "RemoveContainer" containerID="ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec" Oct 10 09:29:30 crc kubenswrapper[4669]: E1010 09:29:30.053223 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec\": container with ID starting with ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec not found: ID does not exist" containerID="ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.053265 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec"} err="failed to get container status \"ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec\": rpc error: code = NotFound desc = could not find container \"ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec\": container with ID starting with ad5a68cebe03493a8059e93059ad40b40f3cb8f6907311cc68156679c90ecaec not found: ID does not exist" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.053290 4669 scope.go:117] "RemoveContainer" containerID="0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4" Oct 10 09:29:30 crc kubenswrapper[4669]: E1010 09:29:30.053754 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4\": container with ID starting with 0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4 not found: ID does not exist" containerID="0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.053777 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4"} err="failed to get container status \"0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4\": rpc error: code = NotFound desc = could not find container \"0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4\": container with ID starting with 0620e76c8e7196db7400a25a7e257b7117cdf7e007a639393f47256ff782b9f4 not found: ID does not exist" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.136281 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-dns-svc\") pod \"eb84008d-41fb-46a0-97d9-4237d4575fe6\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.136391 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm677\" (UniqueName: \"kubernetes.io/projected/eb84008d-41fb-46a0-97d9-4237d4575fe6-kube-api-access-wm677\") pod \"eb84008d-41fb-46a0-97d9-4237d4575fe6\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.136446 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-config\") pod \"eb84008d-41fb-46a0-97d9-4237d4575fe6\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.136477 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-sb\") pod \"eb84008d-41fb-46a0-97d9-4237d4575fe6\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.136511 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-nb\") pod \"eb84008d-41fb-46a0-97d9-4237d4575fe6\" (UID: \"eb84008d-41fb-46a0-97d9-4237d4575fe6\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.148786 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb84008d-41fb-46a0-97d9-4237d4575fe6-kube-api-access-wm677" (OuterVolumeSpecName: "kube-api-access-wm677") pod "eb84008d-41fb-46a0-97d9-4237d4575fe6" (UID: "eb84008d-41fb-46a0-97d9-4237d4575fe6"). InnerVolumeSpecName "kube-api-access-wm677". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.205203 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb84008d-41fb-46a0-97d9-4237d4575fe6" (UID: "eb84008d-41fb-46a0-97d9-4237d4575fe6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.209112 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb84008d-41fb-46a0-97d9-4237d4575fe6" (UID: "eb84008d-41fb-46a0-97d9-4237d4575fe6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.237959 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm677\" (UniqueName: \"kubernetes.io/projected/eb84008d-41fb-46a0-97d9-4237d4575fe6-kube-api-access-wm677\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.237988 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.237996 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.247566 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb84008d-41fb-46a0-97d9-4237d4575fe6" (UID: "eb84008d-41fb-46a0-97d9-4237d4575fe6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.255483 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-config" (OuterVolumeSpecName: "config") pod "eb84008d-41fb-46a0-97d9-4237d4575fe6" (UID: "eb84008d-41fb-46a0-97d9-4237d4575fe6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.339892 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.339926 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb84008d-41fb-46a0-97d9-4237d4575fe6-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.341625 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mgpjg"] Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.347734 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mgpjg"] Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.770646 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.949080 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-run-httpd\") pod \"16462469-a4fc-4197-af0e-f9d4973b6564\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.949165 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-combined-ca-bundle\") pod \"16462469-a4fc-4197-af0e-f9d4973b6564\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.949539 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "16462469-a4fc-4197-af0e-f9d4973b6564" (UID: "16462469-a4fc-4197-af0e-f9d4973b6564"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.949692 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-sg-core-conf-yaml\") pod \"16462469-a4fc-4197-af0e-f9d4973b6564\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.949746 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2kgf\" (UniqueName: \"kubernetes.io/projected/16462469-a4fc-4197-af0e-f9d4973b6564-kube-api-access-w2kgf\") pod \"16462469-a4fc-4197-af0e-f9d4973b6564\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.949767 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-log-httpd\") pod \"16462469-a4fc-4197-af0e-f9d4973b6564\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.949810 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-config-data\") pod \"16462469-a4fc-4197-af0e-f9d4973b6564\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.950076 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "16462469-a4fc-4197-af0e-f9d4973b6564" (UID: "16462469-a4fc-4197-af0e-f9d4973b6564"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.950131 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-scripts\") pod \"16462469-a4fc-4197-af0e-f9d4973b6564\" (UID: \"16462469-a4fc-4197-af0e-f9d4973b6564\") " Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.953781 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.953799 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16462469-a4fc-4197-af0e-f9d4973b6564-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.953872 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16462469-a4fc-4197-af0e-f9d4973b6564-kube-api-access-w2kgf" (OuterVolumeSpecName: "kube-api-access-w2kgf") pod "16462469-a4fc-4197-af0e-f9d4973b6564" (UID: "16462469-a4fc-4197-af0e-f9d4973b6564"). InnerVolumeSpecName "kube-api-access-w2kgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:30 crc kubenswrapper[4669]: I1010 09:29:30.958673 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-scripts" (OuterVolumeSpecName: "scripts") pod "16462469-a4fc-4197-af0e-f9d4973b6564" (UID: "16462469-a4fc-4197-af0e-f9d4973b6564"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.028363 4669 generic.go:334] "Generic (PLEG): container finished" podID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerID="811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba" exitCode=0 Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.028460 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6","Type":"ContainerDied","Data":"811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba"} Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.031844 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "16462469-a4fc-4197-af0e-f9d4973b6564" (UID: "16462469-a4fc-4197-af0e-f9d4973b6564"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.041044 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16462469-a4fc-4197-af0e-f9d4973b6564" (UID: "16462469-a4fc-4197-af0e-f9d4973b6564"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.046107 4669 generic.go:334] "Generic (PLEG): container finished" podID="16462469-a4fc-4197-af0e-f9d4973b6564" containerID="032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46" exitCode=0 Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.046136 4669 generic.go:334] "Generic (PLEG): container finished" podID="16462469-a4fc-4197-af0e-f9d4973b6564" containerID="70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2" exitCode=0 Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.046198 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerDied","Data":"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46"} Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.046225 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerDied","Data":"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2"} Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.046236 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16462469-a4fc-4197-af0e-f9d4973b6564","Type":"ContainerDied","Data":"aaca94c8915ffb5160aa979d95ba136bb3a58454c3e645d73602006498724bf3"} Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.046251 4669 scope.go:117] "RemoveContainer" containerID="c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.046356 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.055243 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.055269 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2kgf\" (UniqueName: \"kubernetes.io/projected/16462469-a4fc-4197-af0e-f9d4973b6564-kube-api-access-w2kgf\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.055278 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.055288 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.079283 4669 scope.go:117] "RemoveContainer" containerID="b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.084251 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-config-data" (OuterVolumeSpecName: "config-data") pod "16462469-a4fc-4197-af0e-f9d4973b6564" (UID: "16462469-a4fc-4197-af0e-f9d4973b6564"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.100649 4669 scope.go:117] "RemoveContainer" containerID="032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.117745 4669 scope.go:117] "RemoveContainer" containerID="70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.140973 4669 scope.go:117] "RemoveContainer" containerID="c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.141342 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71\": container with ID starting with c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71 not found: ID does not exist" containerID="c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.141427 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71"} err="failed to get container status \"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71\": rpc error: code = NotFound desc = could not find container \"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71\": container with ID starting with c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.141517 4669 scope.go:117] "RemoveContainer" containerID="b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.142051 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7\": container with ID starting with b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7 not found: ID does not exist" containerID="b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.142130 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7"} err="failed to get container status \"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7\": rpc error: code = NotFound desc = could not find container \"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7\": container with ID starting with b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.142200 4669 scope.go:117] "RemoveContainer" containerID="032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.142540 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46\": container with ID starting with 032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46 not found: ID does not exist" containerID="032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.142655 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46"} err="failed to get container status \"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46\": rpc error: code = NotFound desc = could not find container \"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46\": container with ID starting with 032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.142741 4669 scope.go:117] "RemoveContainer" containerID="70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.143051 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2\": container with ID starting with 70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2 not found: ID does not exist" containerID="70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.143126 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2"} err="failed to get container status \"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2\": rpc error: code = NotFound desc = could not find container \"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2\": container with ID starting with 70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.143194 4669 scope.go:117] "RemoveContainer" containerID="c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.143436 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71"} err="failed to get container status \"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71\": rpc error: code = NotFound desc = could not find container \"c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71\": container with ID starting with c1a654e93106a7a3fae119b8c8cff7e8d897806f88fcc447c153efe8d7b45b71 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.143520 4669 scope.go:117] "RemoveContainer" containerID="b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.143776 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7"} err="failed to get container status \"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7\": rpc error: code = NotFound desc = could not find container \"b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7\": container with ID starting with b584b3424568d45da1b61dacd246cb177da85a5b3f8795094ef6f2302de05ee7 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.143863 4669 scope.go:117] "RemoveContainer" containerID="032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.144136 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46"} err="failed to get container status \"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46\": rpc error: code = NotFound desc = could not find container \"032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46\": container with ID starting with 032fda151288132114edee1291b44b3876d0982153378e7a534f281862d48e46 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.144218 4669 scope.go:117] "RemoveContainer" containerID="70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.144624 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2"} err="failed to get container status \"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2\": rpc error: code = NotFound desc = could not find container \"70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2\": container with ID starting with 70e48fc47cc5c3d3b2e1dad9e38c472584ee7638c29baadd06efdad258e10af2 not found: ID does not exist" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.156904 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16462469-a4fc-4197-af0e-f9d4973b6564-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.423508 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.429730 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446278 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.446626 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerName="init" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446639 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerName="init" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.446650 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="sg-core" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446655 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="sg-core" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.446666 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-notification-agent" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446671 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-notification-agent" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.446688 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerName="dnsmasq-dns" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446694 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerName="dnsmasq-dns" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.446716 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="proxy-httpd" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446721 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="proxy-httpd" Oct 10 09:29:31 crc kubenswrapper[4669]: E1010 09:29:31.446733 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-central-agent" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446739 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-central-agent" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446884 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-notification-agent" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446893 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="sg-core" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446906 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="proxy-httpd" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446918 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb84008d-41fb-46a0-97d9-4237d4575fe6" containerName="dnsmasq-dns" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.446929 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" containerName="ceilometer-central-agent" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.448253 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.450161 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.467406 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.495991 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.570567 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.570727 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-scripts\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.571141 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-run-httpd\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.571211 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-config-data\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.571271 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gr2d\" (UniqueName: \"kubernetes.io/projected/2bfe642e-726c-4e7d-b376-980946f29671-kube-api-access-7gr2d\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.571637 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.571691 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-log-httpd\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.673539 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.673596 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-scripts\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.673649 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-run-httpd\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.673695 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-config-data\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.673718 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gr2d\" (UniqueName: \"kubernetes.io/projected/2bfe642e-726c-4e7d-b376-980946f29671-kube-api-access-7gr2d\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.673755 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.673777 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-log-httpd\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.674124 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-log-httpd\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.674237 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-run-httpd\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.677694 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.677863 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-scripts\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.679564 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-config-data\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.680603 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.692733 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gr2d\" (UniqueName: \"kubernetes.io/projected/2bfe642e-726c-4e7d-b376-980946f29671-kube-api-access-7gr2d\") pod \"ceilometer-0\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.789278 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.811208 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16462469-a4fc-4197-af0e-f9d4973b6564" path="/var/lib/kubelet/pods/16462469-a4fc-4197-af0e-f9d4973b6564/volumes" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.811886 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb84008d-41fb-46a0-97d9-4237d4575fe6" path="/var/lib/kubelet/pods/eb84008d-41fb-46a0-97d9-4237d4575fe6/volumes" Oct 10 09:29:31 crc kubenswrapper[4669]: I1010 09:29:31.948867 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.073562 4669 generic.go:334] "Generic (PLEG): container finished" podID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerID="1ea8850ed5c4772058ca1f9fe9b9d04a3cb3fb182cd1cfd9bd6c6673f5f00026" exitCode=0 Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.073660 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-588b8ffdfb-2xrm4" event={"ID":"c92807a3-b0ae-48c0-ba7b-1e341aabde2a","Type":"ContainerDied","Data":"1ea8850ed5c4772058ca1f9fe9b9d04a3cb3fb182cd1cfd9bd6c6673f5f00026"} Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.307578 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.502453 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.594869 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-httpd-config\") pod \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.594928 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-config\") pod \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.594967 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hgfm\" (UniqueName: \"kubernetes.io/projected/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-kube-api-access-9hgfm\") pod \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.595027 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-combined-ca-bundle\") pod \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.595059 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-ovndb-tls-certs\") pod \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\" (UID: \"c92807a3-b0ae-48c0-ba7b-1e341aabde2a\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.615357 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c92807a3-b0ae-48c0-ba7b-1e341aabde2a" (UID: "c92807a3-b0ae-48c0-ba7b-1e341aabde2a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.659531 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-kube-api-access-9hgfm" (OuterVolumeSpecName: "kube-api-access-9hgfm") pod "c92807a3-b0ae-48c0-ba7b-1e341aabde2a" (UID: "c92807a3-b0ae-48c0-ba7b-1e341aabde2a"). InnerVolumeSpecName "kube-api-access-9hgfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.684924 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c92807a3-b0ae-48c0-ba7b-1e341aabde2a" (UID: "c92807a3-b0ae-48c0-ba7b-1e341aabde2a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.697550 4669 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.697577 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hgfm\" (UniqueName: \"kubernetes.io/projected/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-kube-api-access-9hgfm\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.697600 4669 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.697955 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c92807a3-b0ae-48c0-ba7b-1e341aabde2a" (UID: "c92807a3-b0ae-48c0-ba7b-1e341aabde2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.705672 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-config" (OuterVolumeSpecName: "config") pod "c92807a3-b0ae-48c0-ba7b-1e341aabde2a" (UID: "c92807a3-b0ae-48c0-ba7b-1e341aabde2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.722560 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.802799 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-combined-ca-bundle\") pod \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.802854 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data\") pod \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.803197 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.803214 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c92807a3-b0ae-48c0-ba7b-1e341aabde2a-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.858657 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" (UID: "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.885541 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data" (OuterVolumeSpecName: "config-data") pod "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" (UID: "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.904478 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data-custom\") pod \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.904567 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-etc-machine-id\") pod \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.904633 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8wgn\" (UniqueName: \"kubernetes.io/projected/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-kube-api-access-c8wgn\") pod \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.904744 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-scripts\") pod \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\" (UID: \"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6\") " Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.904977 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" (UID: "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.905233 4669 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.905257 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.905269 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.907108 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" (UID: "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.907522 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-kube-api-access-c8wgn" (OuterVolumeSpecName: "kube-api-access-c8wgn") pod "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" (UID: "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6"). InnerVolumeSpecName "kube-api-access-c8wgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:32 crc kubenswrapper[4669]: I1010 09:29:32.910169 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-scripts" (OuterVolumeSpecName: "scripts") pod "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" (UID: "ae8ada7a-c97b-423d-b249-bc7c98d8b8b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.007390 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.007433 4669 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.007449 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8wgn\" (UniqueName: \"kubernetes.io/projected/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6-kube-api-access-c8wgn\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.087347 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-588b8ffdfb-2xrm4" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.087478 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-588b8ffdfb-2xrm4" event={"ID":"c92807a3-b0ae-48c0-ba7b-1e341aabde2a","Type":"ContainerDied","Data":"917abb5150a2a61d14e59923222d687d09ba5a7b71d3ae749faf65471d973843"} Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.087536 4669 scope.go:117] "RemoveContainer" containerID="272a322b633e4955e9bedb2be9f7cbb125c4b0428dd4109d2d1ab55e79bdc15a" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.094226 4669 generic.go:334] "Generic (PLEG): container finished" podID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerID="7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305" exitCode=0 Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.094278 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6","Type":"ContainerDied","Data":"7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305"} Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.094304 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8ada7a-c97b-423d-b249-bc7c98d8b8b6","Type":"ContainerDied","Data":"5d3ca8332495dfc2fb09e01a8a8712e6c52db25fc6009c084efb41a649484a33"} Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.094355 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.110244 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerStarted","Data":"3517d41b05febd89b7330a6d8dc2cab091624c50d613c633aac6208fde956370"} Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.134891 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-588b8ffdfb-2xrm4"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.135266 4669 scope.go:117] "RemoveContainer" containerID="1ea8850ed5c4772058ca1f9fe9b9d04a3cb3fb182cd1cfd9bd6c6673f5f00026" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.147069 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-588b8ffdfb-2xrm4"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.157531 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.164072 4669 scope.go:117] "RemoveContainer" containerID="811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.167295 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.173404 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:33 crc kubenswrapper[4669]: E1010 09:29:33.173899 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-api" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.173918 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-api" Oct 10 09:29:33 crc kubenswrapper[4669]: E1010 09:29:33.173933 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-httpd" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.173939 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-httpd" Oct 10 09:29:33 crc kubenswrapper[4669]: E1010 09:29:33.173959 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="cinder-scheduler" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.173966 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="cinder-scheduler" Oct 10 09:29:33 crc kubenswrapper[4669]: E1010 09:29:33.173977 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="probe" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.173983 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="probe" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.174142 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-api" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.174152 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="probe" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.174166 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" containerName="cinder-scheduler" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.174181 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" containerName="neutron-httpd" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.175174 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.180634 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.188496 4669 scope.go:117] "RemoveContainer" containerID="7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.193787 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.215879 4669 scope.go:117] "RemoveContainer" containerID="811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba" Oct 10 09:29:33 crc kubenswrapper[4669]: E1010 09:29:33.216773 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba\": container with ID starting with 811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba not found: ID does not exist" containerID="811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.216809 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba"} err="failed to get container status \"811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba\": rpc error: code = NotFound desc = could not find container \"811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba\": container with ID starting with 811269a8fd0e1070f1b529b7d00a5ef621663ba83382f09605d1e49e2866ffba not found: ID does not exist" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.216841 4669 scope.go:117] "RemoveContainer" containerID="7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305" Oct 10 09:29:33 crc kubenswrapper[4669]: E1010 09:29:33.217306 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305\": container with ID starting with 7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305 not found: ID does not exist" containerID="7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.217331 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305"} err="failed to get container status \"7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305\": rpc error: code = NotFound desc = could not find container \"7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305\": container with ID starting with 7073c43db9eaf40a4ba854b70841f7bdb3361ef73717f1ec038dce7d59b64305 not found: ID does not exist" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.312358 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.312445 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn7xc\" (UniqueName: \"kubernetes.io/projected/23eadb50-c141-42eb-b24c-2a493eec134e-kube-api-access-nn7xc\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.312475 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-config-data\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.312658 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.312783 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23eadb50-c141-42eb-b24c-2a493eec134e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.312834 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-scripts\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.414041 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23eadb50-c141-42eb-b24c-2a493eec134e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.414096 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-scripts\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.414162 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.414179 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23eadb50-c141-42eb-b24c-2a493eec134e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.414207 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn7xc\" (UniqueName: \"kubernetes.io/projected/23eadb50-c141-42eb-b24c-2a493eec134e-kube-api-access-nn7xc\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.414226 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-config-data\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.414253 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.418764 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.419201 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.430533 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-config-data\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.434779 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn7xc\" (UniqueName: \"kubernetes.io/projected/23eadb50-c141-42eb-b24c-2a493eec134e-kube-api-access-nn7xc\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.436704 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23eadb50-c141-42eb-b24c-2a493eec134e-scripts\") pod \"cinder-scheduler-0\" (UID: \"23eadb50-c141-42eb-b24c-2a493eec134e\") " pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.499504 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.605620 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c235-account-create-qclhl"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.617254 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c235-account-create-qclhl" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.620324 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.624747 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c235-account-create-qclhl"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.719562 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6kz5\" (UniqueName: \"kubernetes.io/projected/dbb4d91d-a933-4c7f-bcf9-beab62822018-kube-api-access-s6kz5\") pod \"nova-api-c235-account-create-qclhl\" (UID: \"dbb4d91d-a933-4c7f-bcf9-beab62822018\") " pod="openstack/nova-api-c235-account-create-qclhl" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.815250 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8ada7a-c97b-423d-b249-bc7c98d8b8b6" path="/var/lib/kubelet/pods/ae8ada7a-c97b-423d-b249-bc7c98d8b8b6/volumes" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.816237 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c92807a3-b0ae-48c0-ba7b-1e341aabde2a" path="/var/lib/kubelet/pods/c92807a3-b0ae-48c0-ba7b-1e341aabde2a/volumes" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.816785 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a005-account-create-drmrx"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.820100 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a005-account-create-drmrx" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.821121 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6kz5\" (UniqueName: \"kubernetes.io/projected/dbb4d91d-a933-4c7f-bcf9-beab62822018-kube-api-access-s6kz5\") pod \"nova-api-c235-account-create-qclhl\" (UID: \"dbb4d91d-a933-4c7f-bcf9-beab62822018\") " pod="openstack/nova-api-c235-account-create-qclhl" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.824939 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.829996 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a005-account-create-drmrx"] Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.852521 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6kz5\" (UniqueName: \"kubernetes.io/projected/dbb4d91d-a933-4c7f-bcf9-beab62822018-kube-api-access-s6kz5\") pod \"nova-api-c235-account-create-qclhl\" (UID: \"dbb4d91d-a933-4c7f-bcf9-beab62822018\") " pod="openstack/nova-api-c235-account-create-qclhl" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.922834 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sppch\" (UniqueName: \"kubernetes.io/projected/02fe8a8b-bb2b-4d50-b645-49f4fd745dbd-kube-api-access-sppch\") pod \"nova-cell0-a005-account-create-drmrx\" (UID: \"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd\") " pod="openstack/nova-cell0-a005-account-create-drmrx" Oct 10 09:29:33 crc kubenswrapper[4669]: I1010 09:29:33.975619 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c235-account-create-qclhl" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.025792 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-d5de-account-create-mxqm2"] Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.026364 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sppch\" (UniqueName: \"kubernetes.io/projected/02fe8a8b-bb2b-4d50-b645-49f4fd745dbd-kube-api-access-sppch\") pod \"nova-cell0-a005-account-create-drmrx\" (UID: \"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd\") " pod="openstack/nova-cell0-a005-account-create-drmrx" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.028499 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5de-account-create-mxqm2" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.034570 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.047293 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d5de-account-create-mxqm2"] Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.051705 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sppch\" (UniqueName: \"kubernetes.io/projected/02fe8a8b-bb2b-4d50-b645-49f4fd745dbd-kube-api-access-sppch\") pod \"nova-cell0-a005-account-create-drmrx\" (UID: \"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd\") " pod="openstack/nova-cell0-a005-account-create-drmrx" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.061677 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.130630 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmzck\" (UniqueName: \"kubernetes.io/projected/63afb443-38e7-4f99-98af-007b85175785-kube-api-access-zmzck\") pod \"nova-cell1-d5de-account-create-mxqm2\" (UID: \"63afb443-38e7-4f99-98af-007b85175785\") " pod="openstack/nova-cell1-d5de-account-create-mxqm2" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.146254 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a005-account-create-drmrx" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.146845 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23eadb50-c141-42eb-b24c-2a493eec134e","Type":"ContainerStarted","Data":"f04749e766b1a90cda03920e79519b741717456fcd28f09e71c527e82534ef98"} Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.158200 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerStarted","Data":"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407"} Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.158326 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerStarted","Data":"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50"} Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.233405 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmzck\" (UniqueName: \"kubernetes.io/projected/63afb443-38e7-4f99-98af-007b85175785-kube-api-access-zmzck\") pod \"nova-cell1-d5de-account-create-mxqm2\" (UID: \"63afb443-38e7-4f99-98af-007b85175785\") " pod="openstack/nova-cell1-d5de-account-create-mxqm2" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.269720 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmzck\" (UniqueName: \"kubernetes.io/projected/63afb443-38e7-4f99-98af-007b85175785-kube-api-access-zmzck\") pod \"nova-cell1-d5de-account-create-mxqm2\" (UID: \"63afb443-38e7-4f99-98af-007b85175785\") " pod="openstack/nova-cell1-d5de-account-create-mxqm2" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.398926 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5de-account-create-mxqm2" Oct 10 09:29:34 crc kubenswrapper[4669]: I1010 09:29:34.595853 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c235-account-create-qclhl"] Oct 10 09:29:35 crc kubenswrapper[4669]: I1010 09:29:35.114024 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a005-account-create-drmrx"] Oct 10 09:29:35 crc kubenswrapper[4669]: W1010 09:29:35.125838 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02fe8a8b_bb2b_4d50_b645_49f4fd745dbd.slice/crio-f57eca3eb72afe983977f439b43cf74b4a0d09fdd17446b23aa9af126c811512 WatchSource:0}: Error finding container f57eca3eb72afe983977f439b43cf74b4a0d09fdd17446b23aa9af126c811512: Status 404 returned error can't find the container with id f57eca3eb72afe983977f439b43cf74b4a0d09fdd17446b23aa9af126c811512 Oct 10 09:29:35 crc kubenswrapper[4669]: I1010 09:29:35.173970 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerStarted","Data":"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a"} Oct 10 09:29:35 crc kubenswrapper[4669]: I1010 09:29:35.184869 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c235-account-create-qclhl" event={"ID":"dbb4d91d-a933-4c7f-bcf9-beab62822018","Type":"ContainerStarted","Data":"ea9573f19a0a7de26d8deddf86a4e6be88b99a8f4fcada8865d24c03d0ad7c55"} Oct 10 09:29:35 crc kubenswrapper[4669]: I1010 09:29:35.184912 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c235-account-create-qclhl" event={"ID":"dbb4d91d-a933-4c7f-bcf9-beab62822018","Type":"ContainerStarted","Data":"7955fc4fd3078db8e5febbc01ea1f3e93c92502a908e295c158e68da322a5733"} Oct 10 09:29:35 crc kubenswrapper[4669]: I1010 09:29:35.191738 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a005-account-create-drmrx" event={"ID":"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd","Type":"ContainerStarted","Data":"f57eca3eb72afe983977f439b43cf74b4a0d09fdd17446b23aa9af126c811512"} Oct 10 09:29:35 crc kubenswrapper[4669]: I1010 09:29:35.204078 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c235-account-create-qclhl" podStartSLOduration=2.204060086 podStartE2EDuration="2.204060086s" podCreationTimestamp="2025-10-10 09:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:35.20294207 +0000 UTC m=+1118.218960812" watchObservedRunningTime="2025-10-10 09:29:35.204060086 +0000 UTC m=+1118.220078828" Oct 10 09:29:35 crc kubenswrapper[4669]: I1010 09:29:35.257354 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d5de-account-create-mxqm2"] Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.207995 4669 generic.go:334] "Generic (PLEG): container finished" podID="63afb443-38e7-4f99-98af-007b85175785" containerID="4ecebc4fe8bbce718d521be98fb4e54b676f457f449b6b5f66b255d9db6bce04" exitCode=0 Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.210804 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d5de-account-create-mxqm2" event={"ID":"63afb443-38e7-4f99-98af-007b85175785","Type":"ContainerDied","Data":"4ecebc4fe8bbce718d521be98fb4e54b676f457f449b6b5f66b255d9db6bce04"} Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.210844 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d5de-account-create-mxqm2" event={"ID":"63afb443-38e7-4f99-98af-007b85175785","Type":"ContainerStarted","Data":"b9ba8483b4e2f9498f2fa333eab7899ed8d761f2c473495ddaa83ffd68deeb62"} Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.214102 4669 generic.go:334] "Generic (PLEG): container finished" podID="02fe8a8b-bb2b-4d50-b645-49f4fd745dbd" containerID="24413fed3d637c1706004ffba8e7efda201d191ac13ca1af792d3082b5505611" exitCode=0 Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.214170 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a005-account-create-drmrx" event={"ID":"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd","Type":"ContainerDied","Data":"24413fed3d637c1706004ffba8e7efda201d191ac13ca1af792d3082b5505611"} Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.215664 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23eadb50-c141-42eb-b24c-2a493eec134e","Type":"ContainerStarted","Data":"753727384d281a72db9ed817457aa91bb2470403c57f5ffd8766f57b47096889"} Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.215726 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23eadb50-c141-42eb-b24c-2a493eec134e","Type":"ContainerStarted","Data":"ded4aaeabaf7cb715cf1f3490d1a902eff22d6972e05bcb372a91decc74e5644"} Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.217802 4669 generic.go:334] "Generic (PLEG): container finished" podID="dbb4d91d-a933-4c7f-bcf9-beab62822018" containerID="ea9573f19a0a7de26d8deddf86a4e6be88b99a8f4fcada8865d24c03d0ad7c55" exitCode=0 Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.217830 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c235-account-create-qclhl" event={"ID":"dbb4d91d-a933-4c7f-bcf9-beab62822018","Type":"ContainerDied","Data":"ea9573f19a0a7de26d8deddf86a4e6be88b99a8f4fcada8865d24c03d0ad7c55"} Oct 10 09:29:36 crc kubenswrapper[4669]: I1010 09:29:36.344446 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.344424421 podStartE2EDuration="3.344424421s" podCreationTimestamp="2025-10-10 09:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:29:36.312830245 +0000 UTC m=+1119.328848987" watchObservedRunningTime="2025-10-10 09:29:36.344424421 +0000 UTC m=+1119.360443163" Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.227265 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerStarted","Data":"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc"} Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.227680 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-central-agent" containerID="cri-o://a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" gracePeriod=30 Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.227730 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-notification-agent" containerID="cri-o://9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" gracePeriod=30 Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.227806 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="sg-core" containerID="cri-o://d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" gracePeriod=30 Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.227687 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="proxy-httpd" containerID="cri-o://bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" gracePeriod=30 Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.258876 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.39778461 podStartE2EDuration="6.258855762s" podCreationTimestamp="2025-10-10 09:29:31 +0000 UTC" firstStartedPulling="2025-10-10 09:29:32.369743086 +0000 UTC m=+1115.385761828" lastFinishedPulling="2025-10-10 09:29:36.230814238 +0000 UTC m=+1119.246832980" observedRunningTime="2025-10-10 09:29:37.254247383 +0000 UTC m=+1120.270266125" watchObservedRunningTime="2025-10-10 09:29:37.258855762 +0000 UTC m=+1120.274874504" Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.874983 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5de-account-create-mxqm2" Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.880182 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a005-account-create-drmrx" Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.890677 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c235-account-create-qclhl" Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.908357 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmzck\" (UniqueName: \"kubernetes.io/projected/63afb443-38e7-4f99-98af-007b85175785-kube-api-access-zmzck\") pod \"63afb443-38e7-4f99-98af-007b85175785\" (UID: \"63afb443-38e7-4f99-98af-007b85175785\") " Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.908632 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sppch\" (UniqueName: \"kubernetes.io/projected/02fe8a8b-bb2b-4d50-b645-49f4fd745dbd-kube-api-access-sppch\") pod \"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd\" (UID: \"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd\") " Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.920532 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fe8a8b-bb2b-4d50-b645-49f4fd745dbd-kube-api-access-sppch" (OuterVolumeSpecName: "kube-api-access-sppch") pod "02fe8a8b-bb2b-4d50-b645-49f4fd745dbd" (UID: "02fe8a8b-bb2b-4d50-b645-49f4fd745dbd"). InnerVolumeSpecName "kube-api-access-sppch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:37 crc kubenswrapper[4669]: I1010 09:29:37.928368 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63afb443-38e7-4f99-98af-007b85175785-kube-api-access-zmzck" (OuterVolumeSpecName: "kube-api-access-zmzck") pod "63afb443-38e7-4f99-98af-007b85175785" (UID: "63afb443-38e7-4f99-98af-007b85175785"). InnerVolumeSpecName "kube-api-access-zmzck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.011174 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6kz5\" (UniqueName: \"kubernetes.io/projected/dbb4d91d-a933-4c7f-bcf9-beab62822018-kube-api-access-s6kz5\") pod \"dbb4d91d-a933-4c7f-bcf9-beab62822018\" (UID: \"dbb4d91d-a933-4c7f-bcf9-beab62822018\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.011579 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmzck\" (UniqueName: \"kubernetes.io/projected/63afb443-38e7-4f99-98af-007b85175785-kube-api-access-zmzck\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.011610 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sppch\" (UniqueName: \"kubernetes.io/projected/02fe8a8b-bb2b-4d50-b645-49f4fd745dbd-kube-api-access-sppch\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.017108 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbb4d91d-a933-4c7f-bcf9-beab62822018-kube-api-access-s6kz5" (OuterVolumeSpecName: "kube-api-access-s6kz5") pod "dbb4d91d-a933-4c7f-bcf9-beab62822018" (UID: "dbb4d91d-a933-4c7f-bcf9-beab62822018"). InnerVolumeSpecName "kube-api-access-s6kz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.108992 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.114507 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6kz5\" (UniqueName: \"kubernetes.io/projected/dbb4d91d-a933-4c7f-bcf9-beab62822018-kube-api-access-s6kz5\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.177171 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.216978 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-config-data\") pod \"2bfe642e-726c-4e7d-b376-980946f29671\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.217068 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-sg-core-conf-yaml\") pod \"2bfe642e-726c-4e7d-b376-980946f29671\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.217109 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gr2d\" (UniqueName: \"kubernetes.io/projected/2bfe642e-726c-4e7d-b376-980946f29671-kube-api-access-7gr2d\") pod \"2bfe642e-726c-4e7d-b376-980946f29671\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.217201 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-combined-ca-bundle\") pod \"2bfe642e-726c-4e7d-b376-980946f29671\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.217251 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-log-httpd\") pod \"2bfe642e-726c-4e7d-b376-980946f29671\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.217288 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-scripts\") pod \"2bfe642e-726c-4e7d-b376-980946f29671\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.217347 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-run-httpd\") pod \"2bfe642e-726c-4e7d-b376-980946f29671\" (UID: \"2bfe642e-726c-4e7d-b376-980946f29671\") " Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.219225 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2bfe642e-726c-4e7d-b376-980946f29671" (UID: "2bfe642e-726c-4e7d-b376-980946f29671"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.220017 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2bfe642e-726c-4e7d-b376-980946f29671" (UID: "2bfe642e-726c-4e7d-b376-980946f29671"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.240089 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-scripts" (OuterVolumeSpecName: "scripts") pod "2bfe642e-726c-4e7d-b376-980946f29671" (UID: "2bfe642e-726c-4e7d-b376-980946f29671"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.244225 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bfe642e-726c-4e7d-b376-980946f29671-kube-api-access-7gr2d" (OuterVolumeSpecName: "kube-api-access-7gr2d") pod "2bfe642e-726c-4e7d-b376-980946f29671" (UID: "2bfe642e-726c-4e7d-b376-980946f29671"). InnerVolumeSpecName "kube-api-access-7gr2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265134 4669 generic.go:334] "Generic (PLEG): container finished" podID="2bfe642e-726c-4e7d-b376-980946f29671" containerID="bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" exitCode=0 Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265165 4669 generic.go:334] "Generic (PLEG): container finished" podID="2bfe642e-726c-4e7d-b376-980946f29671" containerID="d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" exitCode=2 Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265172 4669 generic.go:334] "Generic (PLEG): container finished" podID="2bfe642e-726c-4e7d-b376-980946f29671" containerID="9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" exitCode=0 Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265180 4669 generic.go:334] "Generic (PLEG): container finished" podID="2bfe642e-726c-4e7d-b376-980946f29671" containerID="a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" exitCode=0 Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265231 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerDied","Data":"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265256 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerDied","Data":"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265267 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerDied","Data":"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265277 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerDied","Data":"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265285 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2bfe642e-726c-4e7d-b376-980946f29671","Type":"ContainerDied","Data":"3517d41b05febd89b7330a6d8dc2cab091624c50d613c633aac6208fde956370"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265299 4669 scope.go:117] "RemoveContainer" containerID="bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.265425 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.271519 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2bfe642e-726c-4e7d-b376-980946f29671" (UID: "2bfe642e-726c-4e7d-b376-980946f29671"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.285218 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c235-account-create-qclhl" event={"ID":"dbb4d91d-a933-4c7f-bcf9-beab62822018","Type":"ContainerDied","Data":"7955fc4fd3078db8e5febbc01ea1f3e93c92502a908e295c158e68da322a5733"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.285287 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7955fc4fd3078db8e5febbc01ea1f3e93c92502a908e295c158e68da322a5733" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.285366 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c235-account-create-qclhl" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.310539 4669 scope.go:117] "RemoveContainer" containerID="d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.311076 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d5de-account-create-mxqm2" event={"ID":"63afb443-38e7-4f99-98af-007b85175785","Type":"ContainerDied","Data":"b9ba8483b4e2f9498f2fa333eab7899ed8d761f2c473495ddaa83ffd68deeb62"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.311101 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9ba8483b4e2f9498f2fa333eab7899ed8d761f2c473495ddaa83ffd68deeb62" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.311182 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d5de-account-create-mxqm2" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.316427 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a005-account-create-drmrx" event={"ID":"02fe8a8b-bb2b-4d50-b645-49f4fd745dbd","Type":"ContainerDied","Data":"f57eca3eb72afe983977f439b43cf74b4a0d09fdd17446b23aa9af126c811512"} Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.316455 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f57eca3eb72afe983977f439b43cf74b4a0d09fdd17446b23aa9af126c811512" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.316542 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a005-account-create-drmrx" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.324688 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.324731 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gr2d\" (UniqueName: \"kubernetes.io/projected/2bfe642e-726c-4e7d-b376-980946f29671-kube-api-access-7gr2d\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.324743 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.324755 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.324766 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2bfe642e-726c-4e7d-b376-980946f29671-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.356006 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bfe642e-726c-4e7d-b376-980946f29671" (UID: "2bfe642e-726c-4e7d-b376-980946f29671"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.373657 4669 scope.go:117] "RemoveContainer" containerID="9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.407150 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-config-data" (OuterVolumeSpecName: "config-data") pod "2bfe642e-726c-4e7d-b376-980946f29671" (UID: "2bfe642e-726c-4e7d-b376-980946f29671"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.425844 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.425870 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfe642e-726c-4e7d-b376-980946f29671-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.454841 4669 scope.go:117] "RemoveContainer" containerID="a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.500021 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.547011 4669 scope.go:117] "RemoveContainer" containerID="bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.551168 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": container with ID starting with bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc not found: ID does not exist" containerID="bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.551221 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc"} err="failed to get container status \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": rpc error: code = NotFound desc = could not find container \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": container with ID starting with bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.551253 4669 scope.go:117] "RemoveContainer" containerID="d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.551543 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": container with ID starting with d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a not found: ID does not exist" containerID="d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.551576 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a"} err="failed to get container status \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": rpc error: code = NotFound desc = could not find container \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": container with ID starting with d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.551611 4669 scope.go:117] "RemoveContainer" containerID="9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.551848 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": container with ID starting with 9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407 not found: ID does not exist" containerID="9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.551877 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407"} err="failed to get container status \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": rpc error: code = NotFound desc = could not find container \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": container with ID starting with 9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.551924 4669 scope.go:117] "RemoveContainer" containerID="a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.552152 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": container with ID starting with a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50 not found: ID does not exist" containerID="a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.552178 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50"} err="failed to get container status \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": rpc error: code = NotFound desc = could not find container \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": container with ID starting with a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.552197 4669 scope.go:117] "RemoveContainer" containerID="bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.552414 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc"} err="failed to get container status \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": rpc error: code = NotFound desc = could not find container \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": container with ID starting with bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.552438 4669 scope.go:117] "RemoveContainer" containerID="d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.552861 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a"} err="failed to get container status \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": rpc error: code = NotFound desc = could not find container \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": container with ID starting with d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.552887 4669 scope.go:117] "RemoveContainer" containerID="9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.556745 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407"} err="failed to get container status \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": rpc error: code = NotFound desc = could not find container \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": container with ID starting with 9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.556794 4669 scope.go:117] "RemoveContainer" containerID="a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.557327 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50"} err="failed to get container status \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": rpc error: code = NotFound desc = could not find container \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": container with ID starting with a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.557357 4669 scope.go:117] "RemoveContainer" containerID="bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.557553 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc"} err="failed to get container status \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": rpc error: code = NotFound desc = could not find container \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": container with ID starting with bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.557596 4669 scope.go:117] "RemoveContainer" containerID="d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.557843 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a"} err="failed to get container status \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": rpc error: code = NotFound desc = could not find container \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": container with ID starting with d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.557915 4669 scope.go:117] "RemoveContainer" containerID="9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.558269 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407"} err="failed to get container status \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": rpc error: code = NotFound desc = could not find container \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": container with ID starting with 9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.558296 4669 scope.go:117] "RemoveContainer" containerID="a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.558617 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50"} err="failed to get container status \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": rpc error: code = NotFound desc = could not find container \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": container with ID starting with a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.558644 4669 scope.go:117] "RemoveContainer" containerID="bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.558973 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc"} err="failed to get container status \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": rpc error: code = NotFound desc = could not find container \"bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc\": container with ID starting with bde0410c6158f2cca3e00d213809d9cf2e16e498c27f8e9bfbbeacf2468eefdc not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.559002 4669 scope.go:117] "RemoveContainer" containerID="d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.559338 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a"} err="failed to get container status \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": rpc error: code = NotFound desc = could not find container \"d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a\": container with ID starting with d62108dd81fd9584aebf726162169ac6cedea580300acd388b05a9a21297e17a not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.559365 4669 scope.go:117] "RemoveContainer" containerID="9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.559623 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407"} err="failed to get container status \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": rpc error: code = NotFound desc = could not find container \"9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407\": container with ID starting with 9950ce1d7086fb5af45b89ec35450bea214ced1e9faeb1d52e73a1d8a4464407 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.559651 4669 scope.go:117] "RemoveContainer" containerID="a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.560006 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50"} err="failed to get container status \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": rpc error: code = NotFound desc = could not find container \"a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50\": container with ID starting with a8a7db06d141ca0882658c8e7e4e72db2025b2562401ce7c4496d0530f039d50 not found: ID does not exist" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.602116 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.610780 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623422 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.623759 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-central-agent" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623774 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-central-agent" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.623787 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-notification-agent" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623794 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-notification-agent" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.623807 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb4d91d-a933-4c7f-bcf9-beab62822018" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623813 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb4d91d-a933-4c7f-bcf9-beab62822018" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.623826 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63afb443-38e7-4f99-98af-007b85175785" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623831 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="63afb443-38e7-4f99-98af-007b85175785" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.623842 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="proxy-httpd" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623847 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="proxy-httpd" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.623857 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="sg-core" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623862 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="sg-core" Oct 10 09:29:38 crc kubenswrapper[4669]: E1010 09:29:38.623876 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fe8a8b-bb2b-4d50-b645-49f4fd745dbd" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.623901 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fe8a8b-bb2b-4d50-b645-49f4fd745dbd" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.624053 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fe8a8b-bb2b-4d50-b645-49f4fd745dbd" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.624066 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-central-agent" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.624075 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="63afb443-38e7-4f99-98af-007b85175785" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.624083 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="sg-core" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.624090 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="ceilometer-notification-agent" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.624102 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb4d91d-a933-4c7f-bcf9-beab62822018" containerName="mariadb-account-create" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.624110 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfe642e-726c-4e7d-b376-980946f29671" containerName="proxy-httpd" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.625515 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.627469 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.632316 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.640286 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.730096 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-scripts\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.730157 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k46vl\" (UniqueName: \"kubernetes.io/projected/1a805368-2f03-4fe3-8658-ad9cddc6712a-kube-api-access-k46vl\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.730201 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.730297 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-config-data\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.730319 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-log-httpd\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.730344 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.730403 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-run-httpd\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.831989 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.832268 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-run-httpd\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.832399 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-scripts\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.832504 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k46vl\" (UniqueName: \"kubernetes.io/projected/1a805368-2f03-4fe3-8658-ad9cddc6712a-kube-api-access-k46vl\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.832621 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.832730 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-run-httpd\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.832900 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-config-data\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.833023 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-log-httpd\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.833482 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-log-httpd\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.835659 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.836869 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-scripts\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.841736 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.844479 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-config-data\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.852675 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k46vl\" (UniqueName: \"kubernetes.io/projected/1a805368-2f03-4fe3-8658-ad9cddc6712a-kube-api-access-k46vl\") pod \"ceilometer-0\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " pod="openstack/ceilometer-0" Oct 10 09:29:38 crc kubenswrapper[4669]: I1010 09:29:38.949319 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:39 crc kubenswrapper[4669]: I1010 09:29:39.449064 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:39 crc kubenswrapper[4669]: I1010 09:29:39.805049 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bfe642e-726c-4e7d-b376-980946f29671" path="/var/lib/kubelet/pods/2bfe642e-726c-4e7d-b376-980946f29671/volumes" Oct 10 09:29:40 crc kubenswrapper[4669]: I1010 09:29:40.351633 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerStarted","Data":"6ffbb512db5dd9811cd3e49dc36f604e2b55664f85852dc2566433a0b811e804"} Oct 10 09:29:40 crc kubenswrapper[4669]: I1010 09:29:40.351991 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerStarted","Data":"fa11bc6fbcf76ae097e62bb754a595c54e0c171852bbade145c985ef90ece020"} Oct 10 09:29:40 crc kubenswrapper[4669]: I1010 09:29:40.791302 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:41 crc kubenswrapper[4669]: I1010 09:29:41.363119 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerStarted","Data":"5d47a5af6d1976c9d9facb18d02ab2e18fd1caaf27d5ac75b655fa81b0e2c026"} Oct 10 09:29:42 crc kubenswrapper[4669]: I1010 09:29:42.374693 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerStarted","Data":"94344d0012806131e11f4312e1b530b37a4c1598cf1f72d7bde7bcd754da4676"} Oct 10 09:29:43 crc kubenswrapper[4669]: I1010 09:29:43.697073 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.137774 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-c7jmw"] Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.139081 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.141880 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.142705 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.142708 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-d7psv" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.154432 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-c7jmw"] Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.225730 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.226010 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-config-data\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.226146 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-scripts\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.226313 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctffr\" (UniqueName: \"kubernetes.io/projected/12148914-1d9a-49a1-a080-e96fce7f1799-kube-api-access-ctffr\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.328259 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-scripts\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.328664 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctffr\" (UniqueName: \"kubernetes.io/projected/12148914-1d9a-49a1-a080-e96fce7f1799-kube-api-access-ctffr\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.328754 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.328783 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-config-data\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.333220 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.335902 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-scripts\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.343477 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-config-data\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.344602 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctffr\" (UniqueName: \"kubernetes.io/projected/12148914-1d9a-49a1-a080-e96fce7f1799-kube-api-access-ctffr\") pod \"nova-cell0-conductor-db-sync-c7jmw\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.401007 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerStarted","Data":"3602cf4c842d805bd0a9417942f90453ed004b7f64d3546f7e0d39e0f03c0c0e"} Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.401163 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-central-agent" containerID="cri-o://6ffbb512db5dd9811cd3e49dc36f604e2b55664f85852dc2566433a0b811e804" gracePeriod=30 Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.401396 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.401632 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="proxy-httpd" containerID="cri-o://3602cf4c842d805bd0a9417942f90453ed004b7f64d3546f7e0d39e0f03c0c0e" gracePeriod=30 Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.401675 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="sg-core" containerID="cri-o://94344d0012806131e11f4312e1b530b37a4c1598cf1f72d7bde7bcd754da4676" gracePeriod=30 Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.401705 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-notification-agent" containerID="cri-o://5d47a5af6d1976c9d9facb18d02ab2e18fd1caaf27d5ac75b655fa81b0e2c026" gracePeriod=30 Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.430173 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.671059432 podStartE2EDuration="6.430156436s" podCreationTimestamp="2025-10-10 09:29:38 +0000 UTC" firstStartedPulling="2025-10-10 09:29:39.464019353 +0000 UTC m=+1122.480038095" lastFinishedPulling="2025-10-10 09:29:43.223116357 +0000 UTC m=+1126.239135099" observedRunningTime="2025-10-10 09:29:44.423692338 +0000 UTC m=+1127.439711080" watchObservedRunningTime="2025-10-10 09:29:44.430156436 +0000 UTC m=+1127.446175178" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.458398 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:29:44 crc kubenswrapper[4669]: I1010 09:29:44.975829 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-c7jmw"] Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.411502 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" event={"ID":"12148914-1d9a-49a1-a080-e96fce7f1799","Type":"ContainerStarted","Data":"e2a7c56eac8fa1113224171aa86e69eecab99c087a7cdc041320608f35acf950"} Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427187 4669 generic.go:334] "Generic (PLEG): container finished" podID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerID="3602cf4c842d805bd0a9417942f90453ed004b7f64d3546f7e0d39e0f03c0c0e" exitCode=0 Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427395 4669 generic.go:334] "Generic (PLEG): container finished" podID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerID="94344d0012806131e11f4312e1b530b37a4c1598cf1f72d7bde7bcd754da4676" exitCode=2 Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427471 4669 generic.go:334] "Generic (PLEG): container finished" podID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerID="5d47a5af6d1976c9d9facb18d02ab2e18fd1caaf27d5ac75b655fa81b0e2c026" exitCode=0 Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427257 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerDied","Data":"3602cf4c842d805bd0a9417942f90453ed004b7f64d3546f7e0d39e0f03c0c0e"} Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427605 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerDied","Data":"94344d0012806131e11f4312e1b530b37a4c1598cf1f72d7bde7bcd754da4676"} Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427630 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerDied","Data":"5d47a5af6d1976c9d9facb18d02ab2e18fd1caaf27d5ac75b655fa81b0e2c026"} Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427641 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerDied","Data":"6ffbb512db5dd9811cd3e49dc36f604e2b55664f85852dc2566433a0b811e804"} Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427544 4669 generic.go:334] "Generic (PLEG): container finished" podID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerID="6ffbb512db5dd9811cd3e49dc36f604e2b55664f85852dc2566433a0b811e804" exitCode=0 Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427663 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a805368-2f03-4fe3-8658-ad9cddc6712a","Type":"ContainerDied","Data":"fa11bc6fbcf76ae097e62bb754a595c54e0c171852bbade145c985ef90ece020"} Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.427673 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa11bc6fbcf76ae097e62bb754a595c54e0c171852bbade145c985ef90ece020" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.469374 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551380 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-run-httpd\") pod \"1a805368-2f03-4fe3-8658-ad9cddc6712a\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551444 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-log-httpd\") pod \"1a805368-2f03-4fe3-8658-ad9cddc6712a\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551482 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-combined-ca-bundle\") pod \"1a805368-2f03-4fe3-8658-ad9cddc6712a\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551537 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-sg-core-conf-yaml\") pod \"1a805368-2f03-4fe3-8658-ad9cddc6712a\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551810 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-config-data\") pod \"1a805368-2f03-4fe3-8658-ad9cddc6712a\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551889 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k46vl\" (UniqueName: \"kubernetes.io/projected/1a805368-2f03-4fe3-8658-ad9cddc6712a-kube-api-access-k46vl\") pod \"1a805368-2f03-4fe3-8658-ad9cddc6712a\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551981 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-scripts\") pod \"1a805368-2f03-4fe3-8658-ad9cddc6712a\" (UID: \"1a805368-2f03-4fe3-8658-ad9cddc6712a\") " Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.551990 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a805368-2f03-4fe3-8658-ad9cddc6712a" (UID: "1a805368-2f03-4fe3-8658-ad9cddc6712a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.552411 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.552822 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a805368-2f03-4fe3-8658-ad9cddc6712a" (UID: "1a805368-2f03-4fe3-8658-ad9cddc6712a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.568823 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a805368-2f03-4fe3-8658-ad9cddc6712a-kube-api-access-k46vl" (OuterVolumeSpecName: "kube-api-access-k46vl") pod "1a805368-2f03-4fe3-8658-ad9cddc6712a" (UID: "1a805368-2f03-4fe3-8658-ad9cddc6712a"). InnerVolumeSpecName "kube-api-access-k46vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.571577 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-scripts" (OuterVolumeSpecName: "scripts") pod "1a805368-2f03-4fe3-8658-ad9cddc6712a" (UID: "1a805368-2f03-4fe3-8658-ad9cddc6712a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.617860 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a805368-2f03-4fe3-8658-ad9cddc6712a" (UID: "1a805368-2f03-4fe3-8658-ad9cddc6712a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.655748 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.655787 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k46vl\" (UniqueName: \"kubernetes.io/projected/1a805368-2f03-4fe3-8658-ad9cddc6712a-kube-api-access-k46vl\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.655802 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.655817 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a805368-2f03-4fe3-8658-ad9cddc6712a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.697731 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a805368-2f03-4fe3-8658-ad9cddc6712a" (UID: "1a805368-2f03-4fe3-8658-ad9cddc6712a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.757570 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.776334 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-config-data" (OuterVolumeSpecName: "config-data") pod "1a805368-2f03-4fe3-8658-ad9cddc6712a" (UID: "1a805368-2f03-4fe3-8658-ad9cddc6712a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:29:45 crc kubenswrapper[4669]: I1010 09:29:45.859137 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a805368-2f03-4fe3-8658-ad9cddc6712a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.436405 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.471402 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.482468 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496240 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:46 crc kubenswrapper[4669]: E1010 09:29:46.496632 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="proxy-httpd" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496648 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="proxy-httpd" Oct 10 09:29:46 crc kubenswrapper[4669]: E1010 09:29:46.496664 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-notification-agent" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496672 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-notification-agent" Oct 10 09:29:46 crc kubenswrapper[4669]: E1010 09:29:46.496700 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="sg-core" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496707 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="sg-core" Oct 10 09:29:46 crc kubenswrapper[4669]: E1010 09:29:46.496719 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-central-agent" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496724 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-central-agent" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496899 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-central-agent" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496924 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="ceilometer-notification-agent" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496935 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="proxy-httpd" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.496952 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" containerName="sg-core" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.498889 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.509159 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.509164 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.529091 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.576154 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.576204 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-scripts\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.576240 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvxw8\" (UniqueName: \"kubernetes.io/projected/47bc9878-4fc8-4873-a360-c7f62b9df08a-kube-api-access-jvxw8\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.576383 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-log-httpd\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.576428 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-run-httpd\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.576477 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.576509 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-config-data\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.678294 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-run-httpd\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.678699 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.678732 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-config-data\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.678788 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.678814 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-scripts\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.678844 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvxw8\" (UniqueName: \"kubernetes.io/projected/47bc9878-4fc8-4873-a360-c7f62b9df08a-kube-api-access-jvxw8\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.678961 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-log-httpd\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.679453 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-run-httpd\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.679865 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-log-httpd\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.683000 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-scripts\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.684143 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.684367 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.696629 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-config-data\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.701491 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvxw8\" (UniqueName: \"kubernetes.io/projected/47bc9878-4fc8-4873-a360-c7f62b9df08a-kube-api-access-jvxw8\") pod \"ceilometer-0\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " pod="openstack/ceilometer-0" Oct 10 09:29:46 crc kubenswrapper[4669]: I1010 09:29:46.821701 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:29:47 crc kubenswrapper[4669]: I1010 09:29:47.313340 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:47 crc kubenswrapper[4669]: W1010 09:29:47.317147 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47bc9878_4fc8_4873_a360_c7f62b9df08a.slice/crio-c4fa83e431c0900b14a6a693ecb7ecaccced7dfb396115204ce98dfa0e6367e9 WatchSource:0}: Error finding container c4fa83e431c0900b14a6a693ecb7ecaccced7dfb396115204ce98dfa0e6367e9: Status 404 returned error can't find the container with id c4fa83e431c0900b14a6a693ecb7ecaccced7dfb396115204ce98dfa0e6367e9 Oct 10 09:29:47 crc kubenswrapper[4669]: I1010 09:29:47.447038 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerStarted","Data":"c4fa83e431c0900b14a6a693ecb7ecaccced7dfb396115204ce98dfa0e6367e9"} Oct 10 09:29:47 crc kubenswrapper[4669]: I1010 09:29:47.809694 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a805368-2f03-4fe3-8658-ad9cddc6712a" path="/var/lib/kubelet/pods/1a805368-2f03-4fe3-8658-ad9cddc6712a/volumes" Oct 10 09:29:48 crc kubenswrapper[4669]: I1010 09:29:48.463752 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerStarted","Data":"22daae1e3e393adbc5badb65cf109ce65aa85ca70a4e79789d88eb70d290dc32"} Oct 10 09:29:54 crc kubenswrapper[4669]: I1010 09:29:54.275270 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:29:54 crc kubenswrapper[4669]: I1010 09:29:54.275776 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:29:54 crc kubenswrapper[4669]: I1010 09:29:54.401740 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:29:54 crc kubenswrapper[4669]: I1010 09:29:54.523349 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerStarted","Data":"6d41b72a686a80e99200cb40409142fc184ece749c86a950e3a29b113871ab68"} Oct 10 09:29:54 crc kubenswrapper[4669]: I1010 09:29:54.524610 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" event={"ID":"12148914-1d9a-49a1-a080-e96fce7f1799","Type":"ContainerStarted","Data":"8d0bf3c8764456a6044302d994777eef5bda57efc173b6dfed30d103f8783e0e"} Oct 10 09:29:54 crc kubenswrapper[4669]: I1010 09:29:54.541554 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" podStartSLOduration=1.821337936 podStartE2EDuration="10.54153582s" podCreationTimestamp="2025-10-10 09:29:44 +0000 UTC" firstStartedPulling="2025-10-10 09:29:45.003466339 +0000 UTC m=+1128.019485081" lastFinishedPulling="2025-10-10 09:29:53.723664223 +0000 UTC m=+1136.739682965" observedRunningTime="2025-10-10 09:29:54.539158933 +0000 UTC m=+1137.555177665" watchObservedRunningTime="2025-10-10 09:29:54.54153582 +0000 UTC m=+1137.557554552" Oct 10 09:29:55 crc kubenswrapper[4669]: I1010 09:29:55.535146 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerStarted","Data":"1903ecae9e8f9323524e5c20c35f53a326c34aa46a909b98a72659d2fc27520c"} Oct 10 09:29:56 crc kubenswrapper[4669]: I1010 09:29:56.546372 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerStarted","Data":"967f44295992ad51be59ba7fd7522bbdd2ad086ff23bb30404fb11a9f33c4160"} Oct 10 09:29:56 crc kubenswrapper[4669]: I1010 09:29:56.546649 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-central-agent" containerID="cri-o://22daae1e3e393adbc5badb65cf109ce65aa85ca70a4e79789d88eb70d290dc32" gracePeriod=30 Oct 10 09:29:56 crc kubenswrapper[4669]: I1010 09:29:56.546900 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="proxy-httpd" containerID="cri-o://967f44295992ad51be59ba7fd7522bbdd2ad086ff23bb30404fb11a9f33c4160" gracePeriod=30 Oct 10 09:29:56 crc kubenswrapper[4669]: I1010 09:29:56.546998 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="sg-core" containerID="cri-o://1903ecae9e8f9323524e5c20c35f53a326c34aa46a909b98a72659d2fc27520c" gracePeriod=30 Oct 10 09:29:56 crc kubenswrapper[4669]: I1010 09:29:56.547062 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 09:29:56 crc kubenswrapper[4669]: I1010 09:29:56.547095 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-notification-agent" containerID="cri-o://6d41b72a686a80e99200cb40409142fc184ece749c86a950e3a29b113871ab68" gracePeriod=30 Oct 10 09:29:56 crc kubenswrapper[4669]: I1010 09:29:56.568738 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.603743804 podStartE2EDuration="10.568715118s" podCreationTimestamp="2025-10-10 09:29:46 +0000 UTC" firstStartedPulling="2025-10-10 09:29:47.320300931 +0000 UTC m=+1130.336319673" lastFinishedPulling="2025-10-10 09:29:56.285272245 +0000 UTC m=+1139.301290987" observedRunningTime="2025-10-10 09:29:56.565576627 +0000 UTC m=+1139.581595389" watchObservedRunningTime="2025-10-10 09:29:56.568715118 +0000 UTC m=+1139.584733870" Oct 10 09:29:57 crc kubenswrapper[4669]: I1010 09:29:57.583810 4669 generic.go:334] "Generic (PLEG): container finished" podID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerID="1903ecae9e8f9323524e5c20c35f53a326c34aa46a909b98a72659d2fc27520c" exitCode=2 Oct 10 09:29:57 crc kubenswrapper[4669]: I1010 09:29:57.584826 4669 generic.go:334] "Generic (PLEG): container finished" podID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerID="6d41b72a686a80e99200cb40409142fc184ece749c86a950e3a29b113871ab68" exitCode=0 Oct 10 09:29:57 crc kubenswrapper[4669]: I1010 09:29:57.584959 4669 generic.go:334] "Generic (PLEG): container finished" podID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerID="22daae1e3e393adbc5badb65cf109ce65aa85ca70a4e79789d88eb70d290dc32" exitCode=0 Oct 10 09:29:57 crc kubenswrapper[4669]: I1010 09:29:57.585099 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerDied","Data":"1903ecae9e8f9323524e5c20c35f53a326c34aa46a909b98a72659d2fc27520c"} Oct 10 09:29:57 crc kubenswrapper[4669]: I1010 09:29:57.585300 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerDied","Data":"6d41b72a686a80e99200cb40409142fc184ece749c86a950e3a29b113871ab68"} Oct 10 09:29:57 crc kubenswrapper[4669]: I1010 09:29:57.585442 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerDied","Data":"22daae1e3e393adbc5badb65cf109ce65aa85ca70a4e79789d88eb70d290dc32"} Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.170339 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8"] Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.171885 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.182941 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.186844 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.199128 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8"] Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.325363 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-config-volume\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.325803 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-secret-volume\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.325927 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28zxm\" (UniqueName: \"kubernetes.io/projected/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-kube-api-access-28zxm\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.427326 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-config-volume\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.427391 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-secret-volume\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.427478 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28zxm\" (UniqueName: \"kubernetes.io/projected/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-kube-api-access-28zxm\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.428604 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-config-volume\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.441617 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-secret-volume\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.457090 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28zxm\" (UniqueName: \"kubernetes.io/projected/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-kube-api-access-28zxm\") pod \"collect-profiles-29334810-947p8\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.490516 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:00 crc kubenswrapper[4669]: I1010 09:30:00.956303 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8"] Oct 10 09:30:01 crc kubenswrapper[4669]: I1010 09:30:01.620864 4669 generic.go:334] "Generic (PLEG): container finished" podID="a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" containerID="01ba41c0373d6f2a0feb7ebeee3024dc63a9151aaa86eec89df7fa5d3f05e8ff" exitCode=0 Oct 10 09:30:01 crc kubenswrapper[4669]: I1010 09:30:01.621017 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" event={"ID":"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc","Type":"ContainerDied","Data":"01ba41c0373d6f2a0feb7ebeee3024dc63a9151aaa86eec89df7fa5d3f05e8ff"} Oct 10 09:30:01 crc kubenswrapper[4669]: I1010 09:30:01.621171 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" event={"ID":"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc","Type":"ContainerStarted","Data":"3666e72d30e5da97603c9f457d44363d8b30225e1f4f98f233347010aef2256a"} Oct 10 09:30:02 crc kubenswrapper[4669]: I1010 09:30:02.976114 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:02 crc kubenswrapper[4669]: I1010 09:30:02.983800 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28zxm\" (UniqueName: \"kubernetes.io/projected/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-kube-api-access-28zxm\") pod \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " Oct 10 09:30:02 crc kubenswrapper[4669]: I1010 09:30:02.984075 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-config-volume\") pod \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " Oct 10 09:30:02 crc kubenswrapper[4669]: I1010 09:30:02.984278 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-secret-volume\") pod \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\" (UID: \"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc\") " Oct 10 09:30:02 crc kubenswrapper[4669]: I1010 09:30:02.984996 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-config-volume" (OuterVolumeSpecName: "config-volume") pod "a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" (UID: "a00b0f55-ba9f-4cb9-8192-b2db96bc73bc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:30:02 crc kubenswrapper[4669]: I1010 09:30:02.991063 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-kube-api-access-28zxm" (OuterVolumeSpecName: "kube-api-access-28zxm") pod "a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" (UID: "a00b0f55-ba9f-4cb9-8192-b2db96bc73bc"). InnerVolumeSpecName "kube-api-access-28zxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:02 crc kubenswrapper[4669]: I1010 09:30:02.996692 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" (UID: "a00b0f55-ba9f-4cb9-8192-b2db96bc73bc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:03 crc kubenswrapper[4669]: I1010 09:30:03.086618 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:03 crc kubenswrapper[4669]: I1010 09:30:03.086658 4669 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:03 crc kubenswrapper[4669]: I1010 09:30:03.086673 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28zxm\" (UniqueName: \"kubernetes.io/projected/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc-kube-api-access-28zxm\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:03 crc kubenswrapper[4669]: I1010 09:30:03.640420 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" event={"ID":"a00b0f55-ba9f-4cb9-8192-b2db96bc73bc","Type":"ContainerDied","Data":"3666e72d30e5da97603c9f457d44363d8b30225e1f4f98f233347010aef2256a"} Oct 10 09:30:03 crc kubenswrapper[4669]: I1010 09:30:03.640458 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3666e72d30e5da97603c9f457d44363d8b30225e1f4f98f233347010aef2256a" Oct 10 09:30:03 crc kubenswrapper[4669]: I1010 09:30:03.640460 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8" Oct 10 09:30:06 crc kubenswrapper[4669]: I1010 09:30:06.665501 4669 generic.go:334] "Generic (PLEG): container finished" podID="12148914-1d9a-49a1-a080-e96fce7f1799" containerID="8d0bf3c8764456a6044302d994777eef5bda57efc173b6dfed30d103f8783e0e" exitCode=0 Oct 10 09:30:06 crc kubenswrapper[4669]: I1010 09:30:06.665570 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" event={"ID":"12148914-1d9a-49a1-a080-e96fce7f1799","Type":"ContainerDied","Data":"8d0bf3c8764456a6044302d994777eef5bda57efc173b6dfed30d103f8783e0e"} Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.033022 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.181836 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-scripts\") pod \"12148914-1d9a-49a1-a080-e96fce7f1799\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.182988 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctffr\" (UniqueName: \"kubernetes.io/projected/12148914-1d9a-49a1-a080-e96fce7f1799-kube-api-access-ctffr\") pod \"12148914-1d9a-49a1-a080-e96fce7f1799\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.183114 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-config-data\") pod \"12148914-1d9a-49a1-a080-e96fce7f1799\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.183239 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-combined-ca-bundle\") pod \"12148914-1d9a-49a1-a080-e96fce7f1799\" (UID: \"12148914-1d9a-49a1-a080-e96fce7f1799\") " Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.190286 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12148914-1d9a-49a1-a080-e96fce7f1799-kube-api-access-ctffr" (OuterVolumeSpecName: "kube-api-access-ctffr") pod "12148914-1d9a-49a1-a080-e96fce7f1799" (UID: "12148914-1d9a-49a1-a080-e96fce7f1799"). InnerVolumeSpecName "kube-api-access-ctffr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.191641 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-scripts" (OuterVolumeSpecName: "scripts") pod "12148914-1d9a-49a1-a080-e96fce7f1799" (UID: "12148914-1d9a-49a1-a080-e96fce7f1799"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.208963 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-config-data" (OuterVolumeSpecName: "config-data") pod "12148914-1d9a-49a1-a080-e96fce7f1799" (UID: "12148914-1d9a-49a1-a080-e96fce7f1799"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.220355 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12148914-1d9a-49a1-a080-e96fce7f1799" (UID: "12148914-1d9a-49a1-a080-e96fce7f1799"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.285880 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.286299 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctffr\" (UniqueName: \"kubernetes.io/projected/12148914-1d9a-49a1-a080-e96fce7f1799-kube-api-access-ctffr\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.286320 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.286337 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12148914-1d9a-49a1-a080-e96fce7f1799-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.686445 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" event={"ID":"12148914-1d9a-49a1-a080-e96fce7f1799","Type":"ContainerDied","Data":"e2a7c56eac8fa1113224171aa86e69eecab99c087a7cdc041320608f35acf950"} Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.686499 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2a7c56eac8fa1113224171aa86e69eecab99c087a7cdc041320608f35acf950" Oct 10 09:30:08 crc kubenswrapper[4669]: I1010 09:30:08.686577 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-c7jmw" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.012084 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 09:30:09 crc kubenswrapper[4669]: E1010 09:30:09.012517 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" containerName="collect-profiles" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.012538 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" containerName="collect-profiles" Oct 10 09:30:09 crc kubenswrapper[4669]: E1010 09:30:09.012575 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12148914-1d9a-49a1-a080-e96fce7f1799" containerName="nova-cell0-conductor-db-sync" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.012599 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="12148914-1d9a-49a1-a080-e96fce7f1799" containerName="nova-cell0-conductor-db-sync" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.012795 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" containerName="collect-profiles" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.012826 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="12148914-1d9a-49a1-a080-e96fce7f1799" containerName="nova-cell0-conductor-db-sync" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.013475 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.015344 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-d7psv" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.018385 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.040703 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.204741 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82tp6\" (UniqueName: \"kubernetes.io/projected/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-kube-api-access-82tp6\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.204855 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.204875 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.305947 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.306055 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82tp6\" (UniqueName: \"kubernetes.io/projected/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-kube-api-access-82tp6\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.306147 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.310595 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.311627 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.323581 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82tp6\" (UniqueName: \"kubernetes.io/projected/0c654342-12f4-40f3-a4dd-e4e5a76b8b76-kube-api-access-82tp6\") pod \"nova-cell0-conductor-0\" (UID: \"0c654342-12f4-40f3-a4dd-e4e5a76b8b76\") " pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.327647 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:09 crc kubenswrapper[4669]: I1010 09:30:09.769923 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 09:30:10 crc kubenswrapper[4669]: I1010 09:30:10.704391 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0c654342-12f4-40f3-a4dd-e4e5a76b8b76","Type":"ContainerStarted","Data":"d49102fdf1c364d6cda744aa1a57c9b4ee44ee323554470bfd00cbe2af4ac83d"} Oct 10 09:30:10 crc kubenswrapper[4669]: I1010 09:30:10.704813 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:10 crc kubenswrapper[4669]: I1010 09:30:10.704834 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0c654342-12f4-40f3-a4dd-e4e5a76b8b76","Type":"ContainerStarted","Data":"b4661c8c12eaae6f86aa11f09272dfdf54a3b51e077e4a7514ae6bde582517b8"} Oct 10 09:30:14 crc kubenswrapper[4669]: I1010 09:30:14.356764 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 10 09:30:14 crc kubenswrapper[4669]: I1010 09:30:14.385883 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=6.38586052 podStartE2EDuration="6.38586052s" podCreationTimestamp="2025-10-10 09:30:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:10.725392747 +0000 UTC m=+1153.741411589" watchObservedRunningTime="2025-10-10 09:30:14.38586052 +0000 UTC m=+1157.401879262" Oct 10 09:30:14 crc kubenswrapper[4669]: I1010 09:30:14.946151 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-6fl4m"] Oct 10 09:30:14 crc kubenswrapper[4669]: I1010 09:30:14.947659 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:14 crc kubenswrapper[4669]: W1010 09:30:14.949717 4669 reflector.go:561] object-"openstack"/"nova-cell0-manage-config-data": failed to list *v1.Secret: secrets "nova-cell0-manage-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 10 09:30:14 crc kubenswrapper[4669]: E1010 09:30:14.949759 4669 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"nova-cell0-manage-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nova-cell0-manage-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:30:14 crc kubenswrapper[4669]: W1010 09:30:14.949895 4669 reflector.go:561] object-"openstack"/"nova-cell0-manage-scripts": failed to list *v1.Secret: secrets "nova-cell0-manage-scripts" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 10 09:30:14 crc kubenswrapper[4669]: E1010 09:30:14.949955 4669 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"nova-cell0-manage-scripts\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nova-cell0-manage-scripts\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 10 09:30:14 crc kubenswrapper[4669]: I1010 09:30:14.956221 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6fl4m"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.118279 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-config-data\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.118332 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.118383 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn9pj\" (UniqueName: \"kubernetes.io/projected/9c1db2af-8211-4298-8514-047aa3e098cc-kube-api-access-nn9pj\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.118414 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-scripts\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.151472 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.153300 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.155771 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.172318 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.174187 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.181954 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.186676 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.214898 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.242908 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-config-data\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.243281 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.243468 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-config-data\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.243763 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.244759 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.244933 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mplnd\" (UniqueName: \"kubernetes.io/projected/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-kube-api-access-mplnd\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.245054 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn9pj\" (UniqueName: \"kubernetes.io/projected/9c1db2af-8211-4298-8514-047aa3e098cc-kube-api-access-nn9pj\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.245203 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-scripts\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.245548 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.245805 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll2jd\" (UniqueName: \"kubernetes.io/projected/a4422151-89f6-4138-a792-13e1c7473621-kube-api-access-ll2jd\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.254490 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.292749 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.298122 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.300208 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn9pj\" (UniqueName: \"kubernetes.io/projected/9c1db2af-8211-4298-8514-047aa3e098cc-kube-api-access-nn9pj\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.311171 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.341659 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.349465 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.350852 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.352389 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.352559 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz8kx\" (UniqueName: \"kubernetes.io/projected/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-kube-api-access-cz8kx\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.352694 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll2jd\" (UniqueName: \"kubernetes.io/projected/a4422151-89f6-4138-a792-13e1c7473621-kube-api-access-ll2jd\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.352815 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-config-data\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.352940 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.353041 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-logs\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.353197 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.353299 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.353420 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mplnd\" (UniqueName: \"kubernetes.io/projected/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-kube-api-access-mplnd\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.353548 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-config-data\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.360299 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.366198 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-config-data\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.367499 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.373797 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.393786 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.451527 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455469 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-logs\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455536 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455570 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-config-data\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455639 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455657 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pz2k\" (UniqueName: \"kubernetes.io/projected/e9a1e370-0aad-489f-afe0-e6011a4133e8-kube-api-access-6pz2k\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455682 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-config-data\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455714 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz8kx\" (UniqueName: \"kubernetes.io/projected/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-kube-api-access-cz8kx\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.455733 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a1e370-0aad-489f-afe0-e6011a4133e8-logs\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.456256 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-logs\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.464110 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mplnd\" (UniqueName: \"kubernetes.io/projected/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-kube-api-access-mplnd\") pod \"nova-scheduler-0\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.469172 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.469173 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll2jd\" (UniqueName: \"kubernetes.io/projected/a4422151-89f6-4138-a792-13e1c7473621-kube-api-access-ll2jd\") pod \"nova-cell1-novncproxy-0\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.477962 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.486043 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-config-data\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.502964 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.541622 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz8kx\" (UniqueName: \"kubernetes.io/projected/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-kube-api-access-cz8kx\") pod \"nova-api-0\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.571672 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.571731 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pz2k\" (UniqueName: \"kubernetes.io/projected/e9a1e370-0aad-489f-afe0-e6011a4133e8-kube-api-access-6pz2k\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.571778 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-config-data\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.571830 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a1e370-0aad-489f-afe0-e6011a4133e8-logs\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.572369 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a1e370-0aad-489f-afe0-e6011a4133e8-logs\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.579447 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.580131 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-config-data\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.609185 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pz2k\" (UniqueName: \"kubernetes.io/projected/e9a1e370-0aad-489f-afe0-e6011a4133e8-kube-api-access-6pz2k\") pod \"nova-metadata-0\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.628578 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-mbghh"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.630931 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.675619 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrrq8\" (UniqueName: \"kubernetes.io/projected/aab7b93a-d05b-484b-b93a-ada52b09b9d2-kube-api-access-jrrq8\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.680044 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-dns-svc\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.680436 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-nb\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.680636 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-config\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.680734 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-sb\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.681894 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-mbghh"] Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.746691 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.764007 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.782061 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrrq8\" (UniqueName: \"kubernetes.io/projected/aab7b93a-d05b-484b-b93a-ada52b09b9d2-kube-api-access-jrrq8\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.782158 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-dns-svc\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.782217 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-nb\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.782260 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-config\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.782274 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-sb\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.783138 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-sb\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.783776 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-nb\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.783917 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-config\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.785197 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-dns-svc\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.833435 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrrq8\" (UniqueName: \"kubernetes.io/projected/aab7b93a-d05b-484b-b93a-ada52b09b9d2-kube-api-access-jrrq8\") pod \"dnsmasq-dns-54974c8ff5-mbghh\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.870170 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.884899 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-config-data\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:15 crc kubenswrapper[4669]: I1010 09:30:15.987038 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.185730 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.193360 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-scripts\") pod \"nova-cell0-cell-mapping-6fl4m\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.217001 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.480736 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.500776 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.522575 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.669488 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.748413 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8c4xv"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.749716 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.753871 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.753921 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.764791 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8c4xv"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.776473 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5","Type":"ContainerStarted","Data":"e0473e4031a27e95d417b7b0e3a4dabff9750933742d2fbecd8c81df8c78c7f5"} Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.777785 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89","Type":"ContainerStarted","Data":"da954282c7fda84c4d55dc2d2d8857e2832454d70d662f03e57965325e5c7918"} Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.779370 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a4422151-89f6-4138-a792-13e1c7473621","Type":"ContainerStarted","Data":"a8b3d87108e6cac1217cfe6b10c28029ac17aa0f84d2aed8460e97db3e7253ab"} Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.781823 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9a1e370-0aad-489f-afe0-e6011a4133e8","Type":"ContainerStarted","Data":"c2917b9b459e1d69f443d517bc059d4dd91a381c69e4eb78a7d6bbd5a296894b"} Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.808562 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-mbghh"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.816834 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd5vw\" (UniqueName: \"kubernetes.io/projected/789ad8bf-d401-4d16-9752-9ffe5f153fc5-kube-api-access-zd5vw\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.817018 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-scripts\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.817131 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.817230 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-config-data\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.827224 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.919393 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-config-data\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.919500 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd5vw\" (UniqueName: \"kubernetes.io/projected/789ad8bf-d401-4d16-9752-9ffe5f153fc5-kube-api-access-zd5vw\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.919558 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-scripts\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.919605 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.925390 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-config-data\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.931178 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.931322 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-scripts\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.939314 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6fl4m"] Oct 10 09:30:16 crc kubenswrapper[4669]: I1010 09:30:16.950154 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd5vw\" (UniqueName: \"kubernetes.io/projected/789ad8bf-d401-4d16-9752-9ffe5f153fc5-kube-api-access-zd5vw\") pod \"nova-cell1-conductor-db-sync-8c4xv\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.080458 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.601173 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8c4xv"] Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.820391 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" event={"ID":"789ad8bf-d401-4d16-9752-9ffe5f153fc5","Type":"ContainerStarted","Data":"0ad1ce37f1e08678b2ab9114819f2e457c29c3b405a01cd80594b0051586b8b8"} Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.854405 4669 generic.go:334] "Generic (PLEG): container finished" podID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerID="38cd74d82a03a716dbf4a22a88b9523dbf74a938b974d3b831a0ba8bfbdda5c6" exitCode=0 Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.854501 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" event={"ID":"aab7b93a-d05b-484b-b93a-ada52b09b9d2","Type":"ContainerDied","Data":"38cd74d82a03a716dbf4a22a88b9523dbf74a938b974d3b831a0ba8bfbdda5c6"} Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.854527 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" event={"ID":"aab7b93a-d05b-484b-b93a-ada52b09b9d2","Type":"ContainerStarted","Data":"b8fe88eff24a6817a2dad61bc76926577413d866573b25c64c86fe603db9fe28"} Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.875852 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6fl4m" event={"ID":"9c1db2af-8211-4298-8514-047aa3e098cc","Type":"ContainerStarted","Data":"4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b"} Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.875910 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6fl4m" event={"ID":"9c1db2af-8211-4298-8514-047aa3e098cc","Type":"ContainerStarted","Data":"ccf62b661613b4b7d09988f18bdbda578680c6ef1700e777384a21c07d9f8b38"} Oct 10 09:30:17 crc kubenswrapper[4669]: I1010 09:30:17.965986 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-6fl4m" podStartSLOduration=3.965969859 podStartE2EDuration="3.965969859s" podCreationTimestamp="2025-10-10 09:30:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:17.964411089 +0000 UTC m=+1160.980429831" watchObservedRunningTime="2025-10-10 09:30:17.965969859 +0000 UTC m=+1160.981988601" Oct 10 09:30:18 crc kubenswrapper[4669]: I1010 09:30:18.884192 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:18 crc kubenswrapper[4669]: I1010 09:30:18.889193 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" event={"ID":"789ad8bf-d401-4d16-9752-9ffe5f153fc5","Type":"ContainerStarted","Data":"8ed168dfddb073768e6fcdebc0da54e8d6dba95243a136a316671bd552c9b433"} Oct 10 09:30:18 crc kubenswrapper[4669]: I1010 09:30:18.891641 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" event={"ID":"aab7b93a-d05b-484b-b93a-ada52b09b9d2","Type":"ContainerStarted","Data":"f8ed530963ea3feec5e9535df8e231e53a2312c374c2402469b81a4dea85de07"} Oct 10 09:30:18 crc kubenswrapper[4669]: I1010 09:30:18.891769 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:18 crc kubenswrapper[4669]: I1010 09:30:18.896601 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:18 crc kubenswrapper[4669]: I1010 09:30:18.909638 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" podStartSLOduration=2.90961917 podStartE2EDuration="2.90961917s" podCreationTimestamp="2025-10-10 09:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:18.905545458 +0000 UTC m=+1161.921564200" watchObservedRunningTime="2025-10-10 09:30:18.90961917 +0000 UTC m=+1161.925637922" Oct 10 09:30:18 crc kubenswrapper[4669]: I1010 09:30:18.940729 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" podStartSLOduration=3.940707819 podStartE2EDuration="3.940707819s" podCreationTimestamp="2025-10-10 09:30:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:18.93702039 +0000 UTC m=+1161.953039132" watchObservedRunningTime="2025-10-10 09:30:18.940707819 +0000 UTC m=+1161.956726571" Oct 10 09:30:20 crc kubenswrapper[4669]: I1010 09:30:20.916933 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a4422151-89f6-4138-a792-13e1c7473621","Type":"ContainerStarted","Data":"2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f"} Oct 10 09:30:20 crc kubenswrapper[4669]: I1010 09:30:20.918481 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a4422151-89f6-4138-a792-13e1c7473621" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f" gracePeriod=30 Oct 10 09:30:20 crc kubenswrapper[4669]: I1010 09:30:20.921199 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9a1e370-0aad-489f-afe0-e6011a4133e8","Type":"ContainerStarted","Data":"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348"} Oct 10 09:30:20 crc kubenswrapper[4669]: I1010 09:30:20.926516 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5","Type":"ContainerStarted","Data":"1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82"} Oct 10 09:30:20 crc kubenswrapper[4669]: I1010 09:30:20.928234 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89","Type":"ContainerStarted","Data":"9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a"} Oct 10 09:30:20 crc kubenswrapper[4669]: I1010 09:30:20.939362 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.002881753 podStartE2EDuration="5.93934227s" podCreationTimestamp="2025-10-10 09:30:15 +0000 UTC" firstStartedPulling="2025-10-10 09:30:16.530670381 +0000 UTC m=+1159.546689113" lastFinishedPulling="2025-10-10 09:30:20.467130888 +0000 UTC m=+1163.483149630" observedRunningTime="2025-10-10 09:30:20.937898693 +0000 UTC m=+1163.953917435" watchObservedRunningTime="2025-10-10 09:30:20.93934227 +0000 UTC m=+1163.955361002" Oct 10 09:30:20 crc kubenswrapper[4669]: I1010 09:30:20.957712 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.019966112 podStartE2EDuration="5.957681099s" podCreationTimestamp="2025-10-10 09:30:15 +0000 UTC" firstStartedPulling="2025-10-10 09:30:16.528528332 +0000 UTC m=+1159.544547074" lastFinishedPulling="2025-10-10 09:30:20.466243299 +0000 UTC m=+1163.482262061" observedRunningTime="2025-10-10 09:30:20.955131497 +0000 UTC m=+1163.971150239" watchObservedRunningTime="2025-10-10 09:30:20.957681099 +0000 UTC m=+1163.973699841" Oct 10 09:30:21 crc kubenswrapper[4669]: I1010 09:30:21.938950 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5","Type":"ContainerStarted","Data":"7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f"} Oct 10 09:30:21 crc kubenswrapper[4669]: I1010 09:30:21.941250 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9a1e370-0aad-489f-afe0-e6011a4133e8","Type":"ContainerStarted","Data":"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e"} Oct 10 09:30:21 crc kubenswrapper[4669]: I1010 09:30:21.941316 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-log" containerID="cri-o://50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348" gracePeriod=30 Oct 10 09:30:21 crc kubenswrapper[4669]: I1010 09:30:21.941385 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-metadata" containerID="cri-o://e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e" gracePeriod=30 Oct 10 09:30:21 crc kubenswrapper[4669]: I1010 09:30:21.971406 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.035337509 podStartE2EDuration="6.971385622s" podCreationTimestamp="2025-10-10 09:30:15 +0000 UTC" firstStartedPulling="2025-10-10 09:30:16.53655132 +0000 UTC m=+1159.552570062" lastFinishedPulling="2025-10-10 09:30:20.472599433 +0000 UTC m=+1163.488618175" observedRunningTime="2025-10-10 09:30:21.964538253 +0000 UTC m=+1164.980556995" watchObservedRunningTime="2025-10-10 09:30:21.971385622 +0000 UTC m=+1164.987404384" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.494160 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.544221 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-combined-ca-bundle\") pod \"e9a1e370-0aad-489f-afe0-e6011a4133e8\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.544283 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pz2k\" (UniqueName: \"kubernetes.io/projected/e9a1e370-0aad-489f-afe0-e6011a4133e8-kube-api-access-6pz2k\") pod \"e9a1e370-0aad-489f-afe0-e6011a4133e8\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.544331 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-config-data\") pod \"e9a1e370-0aad-489f-afe0-e6011a4133e8\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.544351 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a1e370-0aad-489f-afe0-e6011a4133e8-logs\") pod \"e9a1e370-0aad-489f-afe0-e6011a4133e8\" (UID: \"e9a1e370-0aad-489f-afe0-e6011a4133e8\") " Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.545075 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9a1e370-0aad-489f-afe0-e6011a4133e8-logs" (OuterVolumeSpecName: "logs") pod "e9a1e370-0aad-489f-afe0-e6011a4133e8" (UID: "e9a1e370-0aad-489f-afe0-e6011a4133e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.569771 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a1e370-0aad-489f-afe0-e6011a4133e8-kube-api-access-6pz2k" (OuterVolumeSpecName: "kube-api-access-6pz2k") pod "e9a1e370-0aad-489f-afe0-e6011a4133e8" (UID: "e9a1e370-0aad-489f-afe0-e6011a4133e8"). InnerVolumeSpecName "kube-api-access-6pz2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.584103 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9a1e370-0aad-489f-afe0-e6011a4133e8" (UID: "e9a1e370-0aad-489f-afe0-e6011a4133e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.592928 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-config-data" (OuterVolumeSpecName: "config-data") pod "e9a1e370-0aad-489f-afe0-e6011a4133e8" (UID: "e9a1e370-0aad-489f-afe0-e6011a4133e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.646123 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.646172 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pz2k\" (UniqueName: \"kubernetes.io/projected/e9a1e370-0aad-489f-afe0-e6011a4133e8-kube-api-access-6pz2k\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.646184 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9a1e370-0aad-489f-afe0-e6011a4133e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.646195 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9a1e370-0aad-489f-afe0-e6011a4133e8-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.954271 4669 generic.go:334] "Generic (PLEG): container finished" podID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerID="e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e" exitCode=0 Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.954315 4669 generic.go:334] "Generic (PLEG): container finished" podID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerID="50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348" exitCode=143 Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.955879 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.958726 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9a1e370-0aad-489f-afe0-e6011a4133e8","Type":"ContainerDied","Data":"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e"} Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.958788 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9a1e370-0aad-489f-afe0-e6011a4133e8","Type":"ContainerDied","Data":"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348"} Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.958808 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9a1e370-0aad-489f-afe0-e6011a4133e8","Type":"ContainerDied","Data":"c2917b9b459e1d69f443d517bc059d4dd91a381c69e4eb78a7d6bbd5a296894b"} Oct 10 09:30:22 crc kubenswrapper[4669]: I1010 09:30:22.958836 4669 scope.go:117] "RemoveContainer" containerID="e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.001847 4669 scope.go:117] "RemoveContainer" containerID="50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.017631 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.037337 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.042356 4669 scope.go:117] "RemoveContainer" containerID="e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e" Oct 10 09:30:23 crc kubenswrapper[4669]: E1010 09:30:23.042875 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e\": container with ID starting with e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e not found: ID does not exist" containerID="e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.042926 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e"} err="failed to get container status \"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e\": rpc error: code = NotFound desc = could not find container \"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e\": container with ID starting with e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e not found: ID does not exist" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.042945 4669 scope.go:117] "RemoveContainer" containerID="50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348" Oct 10 09:30:23 crc kubenswrapper[4669]: E1010 09:30:23.043241 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348\": container with ID starting with 50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348 not found: ID does not exist" containerID="50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.043267 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348"} err="failed to get container status \"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348\": rpc error: code = NotFound desc = could not find container \"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348\": container with ID starting with 50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348 not found: ID does not exist" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.043284 4669 scope.go:117] "RemoveContainer" containerID="e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.044163 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e"} err="failed to get container status \"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e\": rpc error: code = NotFound desc = could not find container \"e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e\": container with ID starting with e3e14894eb5f84b2e91472145e4f166a3e9ce4839934450bcbb67d84cb47062e not found: ID does not exist" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.044247 4669 scope.go:117] "RemoveContainer" containerID="50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.044516 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348"} err="failed to get container status \"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348\": rpc error: code = NotFound desc = could not find container \"50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348\": container with ID starting with 50e2fa24ca038f53b96da6d844e826dab5ec371b75843469400bb7e8fff3a348 not found: ID does not exist" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.050858 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:23 crc kubenswrapper[4669]: E1010 09:30:23.051276 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-log" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.051299 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-log" Oct 10 09:30:23 crc kubenswrapper[4669]: E1010 09:30:23.051332 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-metadata" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.051339 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-metadata" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.051508 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-metadata" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.051527 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" containerName="nova-metadata-log" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.058661 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.059800 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.060945 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.061175 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.155784 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-config-data\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.156294 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79wmb\" (UniqueName: \"kubernetes.io/projected/d6090c3a-904a-49ac-b82c-72415fdda3ab-kube-api-access-79wmb\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.156338 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6090c3a-904a-49ac-b82c-72415fdda3ab-logs\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.156382 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.156871 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.258399 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79wmb\" (UniqueName: \"kubernetes.io/projected/d6090c3a-904a-49ac-b82c-72415fdda3ab-kube-api-access-79wmb\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.258713 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6090c3a-904a-49ac-b82c-72415fdda3ab-logs\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.258824 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.259032 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.259126 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-config-data\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.259713 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6090c3a-904a-49ac-b82c-72415fdda3ab-logs\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.264181 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.264579 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-config-data\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.282507 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.293785 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79wmb\" (UniqueName: \"kubernetes.io/projected/d6090c3a-904a-49ac-b82c-72415fdda3ab-kube-api-access-79wmb\") pod \"nova-metadata-0\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.381415 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.807193 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a1e370-0aad-489f-afe0-e6011a4133e8" path="/var/lib/kubelet/pods/e9a1e370-0aad-489f-afe0-e6011a4133e8/volumes" Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.852977 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:23 crc kubenswrapper[4669]: I1010 09:30:23.964386 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6090c3a-904a-49ac-b82c-72415fdda3ab","Type":"ContainerStarted","Data":"4d574d4ba527aeb9e6c99914fb081554fbe4857c0ef3cab9b6f376d9b587181d"} Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.274718 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.275062 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.275385 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.275928 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.275994 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef" gracePeriod=600 Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.979192 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6090c3a-904a-49ac-b82c-72415fdda3ab","Type":"ContainerStarted","Data":"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5"} Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.979662 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6090c3a-904a-49ac-b82c-72415fdda3ab","Type":"ContainerStarted","Data":"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac"} Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.984109 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef" exitCode=0 Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.984148 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef"} Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.984171 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"dcac214f7000ce8e75b0c4db264661af235fd9f2455e77dacca6bf9bb80e4be0"} Oct 10 09:30:24 crc kubenswrapper[4669]: I1010 09:30:24.984192 4669 scope.go:117] "RemoveContainer" containerID="460c2cb0159213ee8a9812e151d381634c493f7eb4c849db7277b1ec7a3b7419" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.015367 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.015340943 podStartE2EDuration="2.015340943s" podCreationTimestamp="2025-10-10 09:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:24.998262503 +0000 UTC m=+1168.014281245" watchObservedRunningTime="2025-10-10 09:30:25.015340943 +0000 UTC m=+1168.031359725" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.479855 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.479924 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.503866 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.506362 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.748174 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.748392 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 09:30:25 crc kubenswrapper[4669]: I1010 09:30:25.989765 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.031501 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.069942 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-df9tj"] Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.070153 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85494b87f-df9tj" podUID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerName="dnsmasq-dns" containerID="cri-o://a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c" gracePeriod=10 Oct 10 09:30:26 crc kubenswrapper[4669]: W1010 09:30:26.609611 4669 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9a1e370_0aad_489f_afe0_e6011a4133e8.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9a1e370_0aad_489f_afe0_e6011a4133e8.slice: no such file or directory Oct 10 09:30:26 crc kubenswrapper[4669]: W1010 09:30:26.625646 4669 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1db2af_8211_4298_8514_047aa3e098cc.slice/crio-conmon-4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1db2af_8211_4298_8514_047aa3e098cc.slice/crio-conmon-4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b.scope: no such file or directory Oct 10 09:30:26 crc kubenswrapper[4669]: W1010 09:30:26.625711 4669 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1db2af_8211_4298_8514_047aa3e098cc.slice/crio-4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1db2af_8211_4298_8514_047aa3e098cc.slice/crio-4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b.scope: no such file or directory Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.752069 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.832461 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.173:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.832647 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.173:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.846425 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-config\") pod \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.846456 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-dns-svc\") pod \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.846518 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-nb\") pod \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.846541 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cclz8\" (UniqueName: \"kubernetes.io/projected/6eb84575-ef26-4a1e-abe1-326f39ecaddf-kube-api-access-cclz8\") pod \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.875071 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eb84575-ef26-4a1e-abe1-326f39ecaddf-kube-api-access-cclz8" (OuterVolumeSpecName: "kube-api-access-cclz8") pod "6eb84575-ef26-4a1e-abe1-326f39ecaddf" (UID: "6eb84575-ef26-4a1e-abe1-326f39ecaddf"). InnerVolumeSpecName "kube-api-access-cclz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.949639 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-sb\") pod \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\" (UID: \"6eb84575-ef26-4a1e-abe1-326f39ecaddf\") " Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.950788 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cclz8\" (UniqueName: \"kubernetes.io/projected/6eb84575-ef26-4a1e-abe1-326f39ecaddf-kube-api-access-cclz8\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.956304 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-config" (OuterVolumeSpecName: "config") pod "6eb84575-ef26-4a1e-abe1-326f39ecaddf" (UID: "6eb84575-ef26-4a1e-abe1-326f39ecaddf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.976406 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6eb84575-ef26-4a1e-abe1-326f39ecaddf" (UID: "6eb84575-ef26-4a1e-abe1-326f39ecaddf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:30:26 crc kubenswrapper[4669]: I1010 09:30:26.989165 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6eb84575-ef26-4a1e-abe1-326f39ecaddf" (UID: "6eb84575-ef26-4a1e-abe1-326f39ecaddf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.012354 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6eb84575-ef26-4a1e-abe1-326f39ecaddf" (UID: "6eb84575-ef26-4a1e-abe1-326f39ecaddf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.022460 4669 generic.go:334] "Generic (PLEG): container finished" podID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerID="967f44295992ad51be59ba7fd7522bbdd2ad086ff23bb30404fb11a9f33c4160" exitCode=137 Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.022507 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerDied","Data":"967f44295992ad51be59ba7fd7522bbdd2ad086ff23bb30404fb11a9f33c4160"} Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.024956 4669 generic.go:334] "Generic (PLEG): container finished" podID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerID="a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c" exitCode=0 Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.025055 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-df9tj" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.025869 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-df9tj" event={"ID":"6eb84575-ef26-4a1e-abe1-326f39ecaddf","Type":"ContainerDied","Data":"a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c"} Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.026025 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-df9tj" event={"ID":"6eb84575-ef26-4a1e-abe1-326f39ecaddf","Type":"ContainerDied","Data":"a73bb82b38507e41ff46d643ace36481a49c312934299a7053d9346985c19227"} Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.026052 4669 scope.go:117] "RemoveContainer" containerID="a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.035194 4669 generic.go:334] "Generic (PLEG): container finished" podID="789ad8bf-d401-4d16-9752-9ffe5f153fc5" containerID="8ed168dfddb073768e6fcdebc0da54e8d6dba95243a136a316671bd552c9b433" exitCode=0 Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.035266 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" event={"ID":"789ad8bf-d401-4d16-9752-9ffe5f153fc5","Type":"ContainerDied","Data":"8ed168dfddb073768e6fcdebc0da54e8d6dba95243a136a316671bd552c9b433"} Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.037169 4669 generic.go:334] "Generic (PLEG): container finished" podID="9c1db2af-8211-4298-8514-047aa3e098cc" containerID="4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b" exitCode=0 Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.037344 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6fl4m" event={"ID":"9c1db2af-8211-4298-8514-047aa3e098cc","Type":"ContainerDied","Data":"4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b"} Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.053081 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.053103 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.053113 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.053124 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6eb84575-ef26-4a1e-abe1-326f39ecaddf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.063653 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.064456 4669 scope.go:117] "RemoveContainer" containerID="df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.090020 4669 scope.go:117] "RemoveContainer" containerID="a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c" Oct 10 09:30:27 crc kubenswrapper[4669]: E1010 09:30:27.090478 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c\": container with ID starting with a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c not found: ID does not exist" containerID="a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.090507 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c"} err="failed to get container status \"a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c\": rpc error: code = NotFound desc = could not find container \"a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c\": container with ID starting with a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c not found: ID does not exist" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.090530 4669 scope.go:117] "RemoveContainer" containerID="df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5" Oct 10 09:30:27 crc kubenswrapper[4669]: E1010 09:30:27.090759 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5\": container with ID starting with df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5 not found: ID does not exist" containerID="df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.090779 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5"} err="failed to get container status \"df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5\": rpc error: code = NotFound desc = could not find container \"df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5\": container with ID starting with df4b460fd8a1a91b1c58a295b98023fab941cff96a60867b41c65f6ca5715cc5 not found: ID does not exist" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.122918 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-df9tj"] Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.137573 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-df9tj"] Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.155083 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvxw8\" (UniqueName: \"kubernetes.io/projected/47bc9878-4fc8-4873-a360-c7f62b9df08a-kube-api-access-jvxw8\") pod \"47bc9878-4fc8-4873-a360-c7f62b9df08a\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.155128 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-combined-ca-bundle\") pod \"47bc9878-4fc8-4873-a360-c7f62b9df08a\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.155186 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-run-httpd\") pod \"47bc9878-4fc8-4873-a360-c7f62b9df08a\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.155977 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "47bc9878-4fc8-4873-a360-c7f62b9df08a" (UID: "47bc9878-4fc8-4873-a360-c7f62b9df08a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.162877 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47bc9878-4fc8-4873-a360-c7f62b9df08a-kube-api-access-jvxw8" (OuterVolumeSpecName: "kube-api-access-jvxw8") pod "47bc9878-4fc8-4873-a360-c7f62b9df08a" (UID: "47bc9878-4fc8-4873-a360-c7f62b9df08a"). InnerVolumeSpecName "kube-api-access-jvxw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.256739 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-sg-core-conf-yaml\") pod \"47bc9878-4fc8-4873-a360-c7f62b9df08a\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.256810 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-log-httpd\") pod \"47bc9878-4fc8-4873-a360-c7f62b9df08a\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.257056 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-scripts\") pod \"47bc9878-4fc8-4873-a360-c7f62b9df08a\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.257082 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-config-data\") pod \"47bc9878-4fc8-4873-a360-c7f62b9df08a\" (UID: \"47bc9878-4fc8-4873-a360-c7f62b9df08a\") " Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.257725 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.257742 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvxw8\" (UniqueName: \"kubernetes.io/projected/47bc9878-4fc8-4873-a360-c7f62b9df08a-kube-api-access-jvxw8\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.257998 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "47bc9878-4fc8-4873-a360-c7f62b9df08a" (UID: "47bc9878-4fc8-4873-a360-c7f62b9df08a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.266923 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-scripts" (OuterVolumeSpecName: "scripts") pod "47bc9878-4fc8-4873-a360-c7f62b9df08a" (UID: "47bc9878-4fc8-4873-a360-c7f62b9df08a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.287850 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47bc9878-4fc8-4873-a360-c7f62b9df08a" (UID: "47bc9878-4fc8-4873-a360-c7f62b9df08a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.289002 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "47bc9878-4fc8-4873-a360-c7f62b9df08a" (UID: "47bc9878-4fc8-4873-a360-c7f62b9df08a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.359541 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.359690 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47bc9878-4fc8-4873-a360-c7f62b9df08a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.360223 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.360338 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.382994 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-config-data" (OuterVolumeSpecName: "config-data") pod "47bc9878-4fc8-4873-a360-c7f62b9df08a" (UID: "47bc9878-4fc8-4873-a360-c7f62b9df08a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.461803 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47bc9878-4fc8-4873-a360-c7f62b9df08a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:27 crc kubenswrapper[4669]: I1010 09:30:27.807029 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" path="/var/lib/kubelet/pods/6eb84575-ef26-4a1e-abe1-326f39ecaddf/volumes" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.049350 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47bc9878-4fc8-4873-a360-c7f62b9df08a","Type":"ContainerDied","Data":"c4fa83e431c0900b14a6a693ecb7ecaccced7dfb396115204ce98dfa0e6367e9"} Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.049382 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.049396 4669 scope.go:117] "RemoveContainer" containerID="967f44295992ad51be59ba7fd7522bbdd2ad086ff23bb30404fb11a9f33c4160" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.083231 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.089465 4669 scope.go:117] "RemoveContainer" containerID="1903ecae9e8f9323524e5c20c35f53a326c34aa46a909b98a72659d2fc27520c" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.094146 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.118100 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:30:28 crc kubenswrapper[4669]: E1010 09:30:28.118838 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-notification-agent" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.118863 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-notification-agent" Oct 10 09:30:28 crc kubenswrapper[4669]: E1010 09:30:28.118901 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="sg-core" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.118915 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="sg-core" Oct 10 09:30:28 crc kubenswrapper[4669]: E1010 09:30:28.118952 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerName="init" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.118959 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerName="init" Oct 10 09:30:28 crc kubenswrapper[4669]: E1010 09:30:28.118983 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-central-agent" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.118991 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-central-agent" Oct 10 09:30:28 crc kubenswrapper[4669]: E1010 09:30:28.119016 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerName="dnsmasq-dns" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.119023 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerName="dnsmasq-dns" Oct 10 09:30:28 crc kubenswrapper[4669]: E1010 09:30:28.119058 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="proxy-httpd" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.119066 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="proxy-httpd" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.119506 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="sg-core" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.119545 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eb84575-ef26-4a1e-abe1-326f39ecaddf" containerName="dnsmasq-dns" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.119564 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-notification-agent" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.119601 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="ceilometer-central-agent" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.119617 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" containerName="proxy-httpd" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.123181 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.140108 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.142045 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.142899 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.162638 4669 scope.go:117] "RemoveContainer" containerID="6d41b72a686a80e99200cb40409142fc184ece749c86a950e3a29b113871ab68" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.180808 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-run-httpd\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.192830 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-config-data\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.192920 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98q64\" (UniqueName: \"kubernetes.io/projected/a0e4e855-9b34-47f0-9e4b-2e21777e7389-kube-api-access-98q64\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.193735 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-log-httpd\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.193864 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.193918 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-scripts\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.193989 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.207770 4669 scope.go:117] "RemoveContainer" containerID="22daae1e3e393adbc5badb65cf109ce65aa85ca70a4e79789d88eb70d290dc32" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.296207 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.296254 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-scripts\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.296293 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.297031 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-run-httpd\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.297056 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-config-data\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.297093 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98q64\" (UniqueName: \"kubernetes.io/projected/a0e4e855-9b34-47f0-9e4b-2e21777e7389-kube-api-access-98q64\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.297143 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-log-httpd\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.297383 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-run-httpd\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.297673 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-log-httpd\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.306897 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.307335 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-config-data\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.314098 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.326297 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98q64\" (UniqueName: \"kubernetes.io/projected/a0e4e855-9b34-47f0-9e4b-2e21777e7389-kube-api-access-98q64\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.337780 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-scripts\") pod \"ceilometer-0\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.382830 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.383034 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.465052 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.543483 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.552879 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.615841 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-config-data\") pod \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.615993 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-scripts\") pod \"9c1db2af-8211-4298-8514-047aa3e098cc\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.616107 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd5vw\" (UniqueName: \"kubernetes.io/projected/789ad8bf-d401-4d16-9752-9ffe5f153fc5-kube-api-access-zd5vw\") pod \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.616143 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn9pj\" (UniqueName: \"kubernetes.io/projected/9c1db2af-8211-4298-8514-047aa3e098cc-kube-api-access-nn9pj\") pod \"9c1db2af-8211-4298-8514-047aa3e098cc\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.616160 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-combined-ca-bundle\") pod \"9c1db2af-8211-4298-8514-047aa3e098cc\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.616200 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-scripts\") pod \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.616233 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-config-data\") pod \"9c1db2af-8211-4298-8514-047aa3e098cc\" (UID: \"9c1db2af-8211-4298-8514-047aa3e098cc\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.616270 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-combined-ca-bundle\") pod \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\" (UID: \"789ad8bf-d401-4d16-9752-9ffe5f153fc5\") " Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.624697 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-scripts" (OuterVolumeSpecName: "scripts") pod "9c1db2af-8211-4298-8514-047aa3e098cc" (UID: "9c1db2af-8211-4298-8514-047aa3e098cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.625128 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c1db2af-8211-4298-8514-047aa3e098cc-kube-api-access-nn9pj" (OuterVolumeSpecName: "kube-api-access-nn9pj") pod "9c1db2af-8211-4298-8514-047aa3e098cc" (UID: "9c1db2af-8211-4298-8514-047aa3e098cc"). InnerVolumeSpecName "kube-api-access-nn9pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.625819 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/789ad8bf-d401-4d16-9752-9ffe5f153fc5-kube-api-access-zd5vw" (OuterVolumeSpecName: "kube-api-access-zd5vw") pod "789ad8bf-d401-4d16-9752-9ffe5f153fc5" (UID: "789ad8bf-d401-4d16-9752-9ffe5f153fc5"). InnerVolumeSpecName "kube-api-access-zd5vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.627344 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-scripts" (OuterVolumeSpecName: "scripts") pod "789ad8bf-d401-4d16-9752-9ffe5f153fc5" (UID: "789ad8bf-d401-4d16-9752-9ffe5f153fc5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.646038 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-config-data" (OuterVolumeSpecName: "config-data") pod "9c1db2af-8211-4298-8514-047aa3e098cc" (UID: "9c1db2af-8211-4298-8514-047aa3e098cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.685931 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "789ad8bf-d401-4d16-9752-9ffe5f153fc5" (UID: "789ad8bf-d401-4d16-9752-9ffe5f153fc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.686478 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c1db2af-8211-4298-8514-047aa3e098cc" (UID: "9c1db2af-8211-4298-8514-047aa3e098cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.686960 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-config-data" (OuterVolumeSpecName: "config-data") pod "789ad8bf-d401-4d16-9752-9ffe5f153fc5" (UID: "789ad8bf-d401-4d16-9752-9ffe5f153fc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718130 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718182 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd5vw\" (UniqueName: \"kubernetes.io/projected/789ad8bf-d401-4d16-9752-9ffe5f153fc5-kube-api-access-zd5vw\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718203 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn9pj\" (UniqueName: \"kubernetes.io/projected/9c1db2af-8211-4298-8514-047aa3e098cc-kube-api-access-nn9pj\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718219 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718234 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718261 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1db2af-8211-4298-8514-047aa3e098cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718276 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:28 crc kubenswrapper[4669]: I1010 09:30:28.718294 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/789ad8bf-d401-4d16-9752-9ffe5f153fc5-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.024055 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.060113 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6fl4m" event={"ID":"9c1db2af-8211-4298-8514-047aa3e098cc","Type":"ContainerDied","Data":"ccf62b661613b4b7d09988f18bdbda578680c6ef1700e777384a21c07d9f8b38"} Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.060147 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccf62b661613b4b7d09988f18bdbda578680c6ef1700e777384a21c07d9f8b38" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.060195 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6fl4m" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.065276 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerStarted","Data":"0fea08f3a08cac29659346bc6791217a4351832771bdb12b92b2c676bd122ff1"} Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.068132 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" event={"ID":"789ad8bf-d401-4d16-9752-9ffe5f153fc5","Type":"ContainerDied","Data":"0ad1ce37f1e08678b2ab9114819f2e457c29c3b405a01cd80594b0051586b8b8"} Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.068173 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ad1ce37f1e08678b2ab9114819f2e457c29c3b405a01cd80594b0051586b8b8" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.068232 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8c4xv" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.133933 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 09:30:29 crc kubenswrapper[4669]: E1010 09:30:29.135307 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="789ad8bf-d401-4d16-9752-9ffe5f153fc5" containerName="nova-cell1-conductor-db-sync" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.135382 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="789ad8bf-d401-4d16-9752-9ffe5f153fc5" containerName="nova-cell1-conductor-db-sync" Oct 10 09:30:29 crc kubenswrapper[4669]: E1010 09:30:29.135461 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c1db2af-8211-4298-8514-047aa3e098cc" containerName="nova-manage" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.135519 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c1db2af-8211-4298-8514-047aa3e098cc" containerName="nova-manage" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.135757 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="789ad8bf-d401-4d16-9752-9ffe5f153fc5" containerName="nova-cell1-conductor-db-sync" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.135824 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c1db2af-8211-4298-8514-047aa3e098cc" containerName="nova-manage" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.136428 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.138700 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.149553 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.226176 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.226462 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.226512 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8hvd\" (UniqueName: \"kubernetes.io/projected/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-kube-api-access-b8hvd\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.253466 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.253758 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-log" containerID="cri-o://1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82" gracePeriod=30 Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.253859 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-api" containerID="cri-o://7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f" gracePeriod=30 Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.277605 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.277848 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" containerName="nova-scheduler-scheduler" containerID="cri-o://9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" gracePeriod=30 Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.316535 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.316813 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-log" containerID="cri-o://acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac" gracePeriod=30 Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.316905 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-metadata" containerID="cri-o://d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5" gracePeriod=30 Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.328108 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.328211 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.328277 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8hvd\" (UniqueName: \"kubernetes.io/projected/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-kube-api-access-b8hvd\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.334424 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.338215 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.356121 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8hvd\" (UniqueName: \"kubernetes.io/projected/9a6e22fc-ac5c-469b-bf5e-3ce1de48968a-kube-api-access-b8hvd\") pod \"nova-cell1-conductor-0\" (UID: \"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a\") " pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.450747 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.826782 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47bc9878-4fc8-4873-a360-c7f62b9df08a" path="/var/lib/kubelet/pods/47bc9878-4fc8-4873-a360-c7f62b9df08a/volumes" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.921960 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:29 crc kubenswrapper[4669]: I1010 09:30:29.954169 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.064848 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-combined-ca-bundle\") pod \"d6090c3a-904a-49ac-b82c-72415fdda3ab\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.065017 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-nova-metadata-tls-certs\") pod \"d6090c3a-904a-49ac-b82c-72415fdda3ab\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.065153 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-config-data\") pod \"d6090c3a-904a-49ac-b82c-72415fdda3ab\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.065187 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79wmb\" (UniqueName: \"kubernetes.io/projected/d6090c3a-904a-49ac-b82c-72415fdda3ab-kube-api-access-79wmb\") pod \"d6090c3a-904a-49ac-b82c-72415fdda3ab\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.065209 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6090c3a-904a-49ac-b82c-72415fdda3ab-logs\") pod \"d6090c3a-904a-49ac-b82c-72415fdda3ab\" (UID: \"d6090c3a-904a-49ac-b82c-72415fdda3ab\") " Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.065960 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6090c3a-904a-49ac-b82c-72415fdda3ab-logs" (OuterVolumeSpecName: "logs") pod "d6090c3a-904a-49ac-b82c-72415fdda3ab" (UID: "d6090c3a-904a-49ac-b82c-72415fdda3ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.111767 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a","Type":"ContainerStarted","Data":"34a432b2f3b7fee64606c1531f0cf0bc9ac2f4b688b2533bfa08ea1f8c9da23d"} Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.126361 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6090c3a-904a-49ac-b82c-72415fdda3ab-kube-api-access-79wmb" (OuterVolumeSpecName: "kube-api-access-79wmb") pod "d6090c3a-904a-49ac-b82c-72415fdda3ab" (UID: "d6090c3a-904a-49ac-b82c-72415fdda3ab"). InnerVolumeSpecName "kube-api-access-79wmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.128772 4669 generic.go:334] "Generic (PLEG): container finished" podID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerID="d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5" exitCode=0 Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.128797 4669 generic.go:334] "Generic (PLEG): container finished" podID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerID="acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac" exitCode=143 Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.128843 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6090c3a-904a-49ac-b82c-72415fdda3ab","Type":"ContainerDied","Data":"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5"} Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.128860 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.128879 4669 scope.go:117] "RemoveContainer" containerID="d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.128868 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6090c3a-904a-49ac-b82c-72415fdda3ab","Type":"ContainerDied","Data":"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac"} Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.129876 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6090c3a-904a-49ac-b82c-72415fdda3ab","Type":"ContainerDied","Data":"4d574d4ba527aeb9e6c99914fb081554fbe4857c0ef3cab9b6f376d9b587181d"} Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.147489 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6090c3a-904a-49ac-b82c-72415fdda3ab" (UID: "d6090c3a-904a-49ac-b82c-72415fdda3ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.167094 4669 scope.go:117] "RemoveContainer" containerID="acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.167393 4669 generic.go:334] "Generic (PLEG): container finished" podID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerID="1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82" exitCode=143 Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.167526 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5","Type":"ContainerDied","Data":"1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82"} Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.167844 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79wmb\" (UniqueName: \"kubernetes.io/projected/d6090c3a-904a-49ac-b82c-72415fdda3ab-kube-api-access-79wmb\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.167869 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6090c3a-904a-49ac-b82c-72415fdda3ab-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.167882 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.183011 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerStarted","Data":"753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b"} Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.267043 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-config-data" (OuterVolumeSpecName: "config-data") pod "d6090c3a-904a-49ac-b82c-72415fdda3ab" (UID: "d6090c3a-904a-49ac-b82c-72415fdda3ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.270352 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.296416 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d6090c3a-904a-49ac-b82c-72415fdda3ab" (UID: "d6090c3a-904a-49ac-b82c-72415fdda3ab"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.372487 4669 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6090c3a-904a-49ac-b82c-72415fdda3ab-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.429081 4669 scope.go:117] "RemoveContainer" containerID="d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5" Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.435859 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5\": container with ID starting with d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5 not found: ID does not exist" containerID="d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.435914 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5"} err="failed to get container status \"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5\": rpc error: code = NotFound desc = could not find container \"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5\": container with ID starting with d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5 not found: ID does not exist" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.435947 4669 scope.go:117] "RemoveContainer" containerID="acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac" Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.436399 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac\": container with ID starting with acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac not found: ID does not exist" containerID="acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.436448 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac"} err="failed to get container status \"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac\": rpc error: code = NotFound desc = could not find container \"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac\": container with ID starting with acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac not found: ID does not exist" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.436476 4669 scope.go:117] "RemoveContainer" containerID="d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.436791 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5"} err="failed to get container status \"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5\": rpc error: code = NotFound desc = could not find container \"d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5\": container with ID starting with d01ece61e4817105d8011dea560645e2a5ec62a5f8386f7e5d9d4e58f49717a5 not found: ID does not exist" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.436814 4669 scope.go:117] "RemoveContainer" containerID="acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.437035 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac"} err="failed to get container status \"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac\": rpc error: code = NotFound desc = could not find container \"acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac\": container with ID starting with acdceb6dcb1709d1e921fa1a93e7f50b66e36e0d778409872e5a7f3d58d372ac not found: ID does not exist" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.473211 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.499501 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.502513 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.507050 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.507126 4669 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" containerName="nova-scheduler-scheduler" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.510071 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.520389 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.520885 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-log" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.520912 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-log" Oct 10 09:30:30 crc kubenswrapper[4669]: E1010 09:30:30.520927 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-metadata" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.520936 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-metadata" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.521126 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-log" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.521165 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" containerName="nova-metadata-metadata" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.522349 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.526991 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.558613 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.559294 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.589314 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.589441 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwjrb\" (UniqueName: \"kubernetes.io/projected/93f92d98-29f6-44b2-9087-1dbd9d74aa62-kube-api-access-kwjrb\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.589474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-config-data\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.589504 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93f92d98-29f6-44b2-9087-1dbd9d74aa62-logs\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.589534 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.691464 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwjrb\" (UniqueName: \"kubernetes.io/projected/93f92d98-29f6-44b2-9087-1dbd9d74aa62-kube-api-access-kwjrb\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.691516 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-config-data\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.691540 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93f92d98-29f6-44b2-9087-1dbd9d74aa62-logs\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.691577 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.691669 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.692066 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93f92d98-29f6-44b2-9087-1dbd9d74aa62-logs\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.698104 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-config-data\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.701148 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.704775 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.710421 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwjrb\" (UniqueName: \"kubernetes.io/projected/93f92d98-29f6-44b2-9087-1dbd9d74aa62-kube-api-access-kwjrb\") pod \"nova-metadata-0\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " pod="openstack/nova-metadata-0" Oct 10 09:30:30 crc kubenswrapper[4669]: I1010 09:30:30.873875 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:30:31 crc kubenswrapper[4669]: I1010 09:30:31.245833 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerStarted","Data":"895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336"} Oct 10 09:30:31 crc kubenswrapper[4669]: I1010 09:30:31.252695 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9a6e22fc-ac5c-469b-bf5e-3ce1de48968a","Type":"ContainerStarted","Data":"9aeda69473d12de4bfe540403d654589b665d424f4c29c150a7d54c8d0cfa679"} Oct 10 09:30:31 crc kubenswrapper[4669]: I1010 09:30:31.252926 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:31 crc kubenswrapper[4669]: I1010 09:30:31.270010 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.269994564 podStartE2EDuration="2.269994564s" podCreationTimestamp="2025-10-10 09:30:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:31.267211575 +0000 UTC m=+1174.283230317" watchObservedRunningTime="2025-10-10 09:30:31.269994564 +0000 UTC m=+1174.286013306" Oct 10 09:30:31 crc kubenswrapper[4669]: I1010 09:30:31.385417 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:30:31 crc kubenswrapper[4669]: I1010 09:30:31.806993 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6090c3a-904a-49ac-b82c-72415fdda3ab" path="/var/lib/kubelet/pods/d6090c3a-904a-49ac-b82c-72415fdda3ab/volumes" Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.270450 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerStarted","Data":"b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06"} Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.272466 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93f92d98-29f6-44b2-9087-1dbd9d74aa62","Type":"ContainerStarted","Data":"cf98206075bede22a040c201cc3d0a2111f3715fc14c2ba83e06b94d116d4a43"} Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.272503 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93f92d98-29f6-44b2-9087-1dbd9d74aa62","Type":"ContainerStarted","Data":"202bf293457faa36e046bce932af6e750f2bbb5c297dfe57f8d63ffc21b6e0be"} Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.272516 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93f92d98-29f6-44b2-9087-1dbd9d74aa62","Type":"ContainerStarted","Data":"82137489e3dd19ea7f930324656e7c39f29afaa23ad813e997e404c1fbce341d"} Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.868886 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.887153 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.887132779 podStartE2EDuration="2.887132779s" podCreationTimestamp="2025-10-10 09:30:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:32.301446117 +0000 UTC m=+1175.317464869" watchObservedRunningTime="2025-10-10 09:30:32.887132779 +0000 UTC m=+1175.903151521" Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.933720 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-combined-ca-bundle\") pod \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.934900 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-config-data\") pod \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.934946 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz8kx\" (UniqueName: \"kubernetes.io/projected/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-kube-api-access-cz8kx\") pod \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.935030 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-logs\") pod \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\" (UID: \"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5\") " Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.935797 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-logs" (OuterVolumeSpecName: "logs") pod "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" (UID: "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.937777 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.940861 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-kube-api-access-cz8kx" (OuterVolumeSpecName: "kube-api-access-cz8kx") pod "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" (UID: "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5"). InnerVolumeSpecName "kube-api-access-cz8kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.966523 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" (UID: "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:32 crc kubenswrapper[4669]: I1010 09:30:32.968786 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-config-data" (OuterVolumeSpecName: "config-data") pod "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" (UID: "b40dac32-8ce7-40fe-bbd5-25e67b9efdf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.041564 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.041617 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.041628 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz8kx\" (UniqueName: \"kubernetes.io/projected/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5-kube-api-access-cz8kx\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.067466 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.143124 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-combined-ca-bundle\") pod \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.143609 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-config-data\") pod \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.143666 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mplnd\" (UniqueName: \"kubernetes.io/projected/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-kube-api-access-mplnd\") pod \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\" (UID: \"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89\") " Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.149687 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-kube-api-access-mplnd" (OuterVolumeSpecName: "kube-api-access-mplnd") pod "cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" (UID: "cbcd4c95-48d2-405e-af12-2b6ff9fe1b89"). InnerVolumeSpecName "kube-api-access-mplnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.171350 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" (UID: "cbcd4c95-48d2-405e-af12-2b6ff9fe1b89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.177573 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-config-data" (OuterVolumeSpecName: "config-data") pod "cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" (UID: "cbcd4c95-48d2-405e-af12-2b6ff9fe1b89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.246111 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.246144 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.246154 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mplnd\" (UniqueName: \"kubernetes.io/projected/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89-kube-api-access-mplnd\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.283025 4669 generic.go:334] "Generic (PLEG): container finished" podID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerID="7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f" exitCode=0 Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.283079 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5","Type":"ContainerDied","Data":"7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f"} Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.283135 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b40dac32-8ce7-40fe-bbd5-25e67b9efdf5","Type":"ContainerDied","Data":"e0473e4031a27e95d417b7b0e3a4dabff9750933742d2fbecd8c81df8c78c7f5"} Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.283156 4669 scope.go:117] "RemoveContainer" containerID="7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.284136 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.285162 4669 generic.go:334] "Generic (PLEG): container finished" podID="cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" containerID="9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" exitCode=0 Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.286216 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.289385 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89","Type":"ContainerDied","Data":"9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a"} Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.289436 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbcd4c95-48d2-405e-af12-2b6ff9fe1b89","Type":"ContainerDied","Data":"da954282c7fda84c4d55dc2d2d8857e2832454d70d662f03e57965325e5c7918"} Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.312337 4669 scope.go:117] "RemoveContainer" containerID="1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.330569 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.341430 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.344786 4669 scope.go:117] "RemoveContainer" containerID="7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f" Oct 10 09:30:33 crc kubenswrapper[4669]: E1010 09:30:33.353720 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f\": container with ID starting with 7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f not found: ID does not exist" containerID="7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.353764 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f"} err="failed to get container status \"7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f\": rpc error: code = NotFound desc = could not find container \"7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f\": container with ID starting with 7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f not found: ID does not exist" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.353791 4669 scope.go:117] "RemoveContainer" containerID="1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.357184 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: E1010 09:30:33.361041 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82\": container with ID starting with 1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82 not found: ID does not exist" containerID="1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.361081 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82"} err="failed to get container status \"1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82\": rpc error: code = NotFound desc = could not find container \"1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82\": container with ID starting with 1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82 not found: ID does not exist" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.361107 4669 scope.go:117] "RemoveContainer" containerID="9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.378317 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.397315 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: E1010 09:30:33.397996 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-api" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.398017 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-api" Oct 10 09:30:33 crc kubenswrapper[4669]: E1010 09:30:33.398070 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-log" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.398080 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-log" Oct 10 09:30:33 crc kubenswrapper[4669]: E1010 09:30:33.398091 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" containerName="nova-scheduler-scheduler" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.398100 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" containerName="nova-scheduler-scheduler" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.398492 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-log" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.398713 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" containerName="nova-api-api" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.398738 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" containerName="nova-scheduler-scheduler" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.400296 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.411144 4669 scope.go:117] "RemoveContainer" containerID="9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" Oct 10 09:30:33 crc kubenswrapper[4669]: E1010 09:30:33.413009 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a\": container with ID starting with 9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a not found: ID does not exist" containerID="9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.413048 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a"} err="failed to get container status \"9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a\": rpc error: code = NotFound desc = could not find container \"9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a\": container with ID starting with 9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a not found: ID does not exist" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.416895 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.417670 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.419551 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.422237 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.438325 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.451141 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fctg\" (UniqueName: \"kubernetes.io/projected/782e0488-3fea-40ae-aeff-766abc1abe9b-kube-api-access-4fctg\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.451264 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.451296 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/782e0488-3fea-40ae-aeff-766abc1abe9b-logs\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.451568 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-config-data\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.457214 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.554029 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-config-data\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.554377 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fctg\" (UniqueName: \"kubernetes.io/projected/782e0488-3fea-40ae-aeff-766abc1abe9b-kube-api-access-4fctg\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.554499 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-config-data\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.554863 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.554989 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f67wt\" (UniqueName: \"kubernetes.io/projected/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-kube-api-access-f67wt\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.555093 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.555179 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/782e0488-3fea-40ae-aeff-766abc1abe9b-logs\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.555781 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/782e0488-3fea-40ae-aeff-766abc1abe9b-logs\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.558562 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.559101 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-config-data\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.575819 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fctg\" (UniqueName: \"kubernetes.io/projected/782e0488-3fea-40ae-aeff-766abc1abe9b-kube-api-access-4fctg\") pod \"nova-api-0\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.657345 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-config-data\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.657431 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.657453 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f67wt\" (UniqueName: \"kubernetes.io/projected/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-kube-api-access-f67wt\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.662298 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-config-data\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.662757 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.679887 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f67wt\" (UniqueName: \"kubernetes.io/projected/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-kube-api-access-f67wt\") pod \"nova-scheduler-0\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.755633 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.768136 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.810069 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40dac32-8ce7-40fe-bbd5-25e67b9efdf5" path="/var/lib/kubelet/pods/b40dac32-8ce7-40fe-bbd5-25e67b9efdf5/volumes" Oct 10 09:30:33 crc kubenswrapper[4669]: I1010 09:30:33.810969 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbcd4c95-48d2-405e-af12-2b6ff9fe1b89" path="/var/lib/kubelet/pods/cbcd4c95-48d2-405e-af12-2b6ff9fe1b89/volumes" Oct 10 09:30:34 crc kubenswrapper[4669]: I1010 09:30:34.234222 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:34 crc kubenswrapper[4669]: I1010 09:30:34.319480 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"782e0488-3fea-40ae-aeff-766abc1abe9b","Type":"ContainerStarted","Data":"d64019728430070b43e27c1eb6b41bf08e39c8d551b8a94d296ba8978e36f658"} Oct 10 09:30:34 crc kubenswrapper[4669]: I1010 09:30:34.322776 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:30:34 crc kubenswrapper[4669]: I1010 09:30:34.324956 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerStarted","Data":"1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70"} Oct 10 09:30:34 crc kubenswrapper[4669]: I1010 09:30:34.325299 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 09:30:34 crc kubenswrapper[4669]: I1010 09:30:34.362339 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.318292414 podStartE2EDuration="6.362318489s" podCreationTimestamp="2025-10-10 09:30:28 +0000 UTC" firstStartedPulling="2025-10-10 09:30:29.02990094 +0000 UTC m=+1172.045919692" lastFinishedPulling="2025-10-10 09:30:33.073927025 +0000 UTC m=+1176.089945767" observedRunningTime="2025-10-10 09:30:34.355285664 +0000 UTC m=+1177.371304396" watchObservedRunningTime="2025-10-10 09:30:34.362318489 +0000 UTC m=+1177.378337231" Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.356141 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"782e0488-3fea-40ae-aeff-766abc1abe9b","Type":"ContainerStarted","Data":"f4f0191739f96568b968ad3f4831179a17641b91096da7498509d357f8a74e27"} Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.356488 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"782e0488-3fea-40ae-aeff-766abc1abe9b","Type":"ContainerStarted","Data":"4a0c2697855a129a2341d961950275033c11431bede8136d79befbbf4782fde3"} Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.361861 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e","Type":"ContainerStarted","Data":"8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254"} Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.361941 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e","Type":"ContainerStarted","Data":"e4c1d22b2a5e002eb55abb1ccae5e28fd9a1f556b6d51cac93ed98deba04507b"} Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.387061 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.387036686 podStartE2EDuration="2.387036686s" podCreationTimestamp="2025-10-10 09:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:35.378512042 +0000 UTC m=+1178.394530794" watchObservedRunningTime="2025-10-10 09:30:35.387036686 +0000 UTC m=+1178.403055428" Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.411447 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.411427741 podStartE2EDuration="2.411427741s" podCreationTimestamp="2025-10-10 09:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:35.398366781 +0000 UTC m=+1178.414385523" watchObservedRunningTime="2025-10-10 09:30:35.411427741 +0000 UTC m=+1178.427446483" Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.876011 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 09:30:35 crc kubenswrapper[4669]: I1010 09:30:35.876401 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 09:30:38 crc kubenswrapper[4669]: I1010 09:30:38.769169 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 09:30:39 crc kubenswrapper[4669]: I1010 09:30:39.491260 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 10 09:30:40 crc kubenswrapper[4669]: I1010 09:30:40.876157 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 09:30:40 crc kubenswrapper[4669]: I1010 09:30:40.876439 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 09:30:41 crc kubenswrapper[4669]: I1010 09:30:41.887804 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 09:30:41 crc kubenswrapper[4669]: I1010 09:30:41.887829 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 09:30:43 crc kubenswrapper[4669]: I1010 09:30:43.756252 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 09:30:43 crc kubenswrapper[4669]: I1010 09:30:43.756633 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 09:30:43 crc kubenswrapper[4669]: I1010 09:30:43.768623 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 09:30:43 crc kubenswrapper[4669]: I1010 09:30:43.806448 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 09:30:44 crc kubenswrapper[4669]: I1010 09:30:44.472953 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 09:30:44 crc kubenswrapper[4669]: I1010 09:30:44.838907 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 09:30:44 crc kubenswrapper[4669]: I1010 09:30:44.838948 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 09:30:50 crc kubenswrapper[4669]: I1010 09:30:50.885647 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 09:30:50 crc kubenswrapper[4669]: I1010 09:30:50.889114 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 09:30:50 crc kubenswrapper[4669]: I1010 09:30:50.901155 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 09:30:50 crc kubenswrapper[4669]: W1010 09:30:50.977326 4669 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbcd4c95_48d2_405e_af12_2b6ff9fe1b89.slice/crio-9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbcd4c95_48d2_405e_af12_2b6ff9fe1b89.slice/crio-9db0747fcd20869d2818b1e2d670cb22dfd6badcdb5066a6c6f008c08c72ad0a.scope: no such file or directory Oct 10 09:30:50 crc kubenswrapper[4669]: W1010 09:30:50.977656 4669 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6090c3a_904a_49ac_b82c_72415fdda3ab.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6090c3a_904a_49ac_b82c_72415fdda3ab.slice: no such file or directory Oct 10 09:30:50 crc kubenswrapper[4669]: W1010 09:30:50.982828 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40dac32_8ce7_40fe_bbd5_25e67b9efdf5.slice/crio-1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82.scope WatchSource:0}: Error finding container 1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82: Status 404 returned error can't find the container with id 1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82 Oct 10 09:30:50 crc kubenswrapper[4669]: W1010 09:30:50.984756 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40dac32_8ce7_40fe_bbd5_25e67b9efdf5.slice/crio-7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f.scope WatchSource:0}: Error finding container 7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f: Status 404 returned error can't find the container with id 7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f Oct 10 09:30:51 crc kubenswrapper[4669]: E1010 09:30:51.227777 4669 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40dac32_8ce7_40fe_bbd5_25e67b9efdf5.slice/crio-conmon-1da830124e84dd95ea086989719ce1285ae62532671b885424e2cb1584bfbf82.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40dac32_8ce7_40fe_bbd5_25e67b9efdf5.slice/crio-e0473e4031a27e95d417b7b0e3a4dabff9750933742d2fbecd8c81df8c78c7f5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47bc9878_4fc8_4873_a360_c7f62b9df08a.slice/crio-c4fa83e431c0900b14a6a693ecb7ecaccced7dfb396115204ce98dfa0e6367e9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb84575_ef26_4a1e_abe1_326f39ecaddf.slice/crio-a73bb82b38507e41ff46d643ace36481a49c312934299a7053d9346985c19227\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47bc9878_4fc8_4873_a360_c7f62b9df08a.slice/crio-conmon-967f44295992ad51be59ba7fd7522bbdd2ad086ff23bb30404fb11a9f33c4160.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47bc9878_4fc8_4873_a360_c7f62b9df08a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaddb758f_1f34_4793_af67_1a54167543b9.slice/crio-d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47bc9878_4fc8_4873_a360_c7f62b9df08a.slice/crio-967f44295992ad51be59ba7fd7522bbdd2ad086ff23bb30404fb11a9f33c4160.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb84575_ef26_4a1e_abe1_326f39ecaddf.slice/crio-conmon-a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaddb758f_1f34_4793_af67_1a54167543b9.slice/crio-conmon-d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb84575_ef26_4a1e_abe1_326f39ecaddf.slice/crio-a1f658e701ff45671c741f06aaaba4b3c0cd587df4b911d4c78df4e8262b046c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1db2af_8211_4298_8514_047aa3e098cc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4422151_89f6_4138_a792_13e1c7473621.slice/crio-conmon-2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40dac32_8ce7_40fe_bbd5_25e67b9efdf5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40dac32_8ce7_40fe_bbd5_25e67b9efdf5.slice/crio-conmon-7da2b61f33c9ab19262f4aab69ab3f7f0ceaf4ca47bbc71639f7c9c5dde0d62f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb84575_ef26_4a1e_abe1_326f39ecaddf.slice\": RecentStats: unable to find data in memory cache]" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.364003 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.489258 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll2jd\" (UniqueName: \"kubernetes.io/projected/a4422151-89f6-4138-a792-13e1c7473621-kube-api-access-ll2jd\") pod \"a4422151-89f6-4138-a792-13e1c7473621\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.489308 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-config-data\") pod \"a4422151-89f6-4138-a792-13e1c7473621\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.489340 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-combined-ca-bundle\") pod \"a4422151-89f6-4138-a792-13e1c7473621\" (UID: \"a4422151-89f6-4138-a792-13e1c7473621\") " Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.495642 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4422151-89f6-4138-a792-13e1c7473621-kube-api-access-ll2jd" (OuterVolumeSpecName: "kube-api-access-ll2jd") pod "a4422151-89f6-4138-a792-13e1c7473621" (UID: "a4422151-89f6-4138-a792-13e1c7473621"). InnerVolumeSpecName "kube-api-access-ll2jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.524958 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-config-data" (OuterVolumeSpecName: "config-data") pod "a4422151-89f6-4138-a792-13e1c7473621" (UID: "a4422151-89f6-4138-a792-13e1c7473621"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.525512 4669 generic.go:334] "Generic (PLEG): container finished" podID="a4422151-89f6-4138-a792-13e1c7473621" containerID="2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f" exitCode=137 Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.525565 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.525597 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a4422151-89f6-4138-a792-13e1c7473621","Type":"ContainerDied","Data":"2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f"} Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.525659 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a4422151-89f6-4138-a792-13e1c7473621","Type":"ContainerDied","Data":"a8b3d87108e6cac1217cfe6b10c28029ac17aa0f84d2aed8460e97db3e7253ab"} Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.525681 4669 scope.go:117] "RemoveContainer" containerID="2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.527391 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4422151-89f6-4138-a792-13e1c7473621" (UID: "a4422151-89f6-4138-a792-13e1c7473621"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.535553 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.593716 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll2jd\" (UniqueName: \"kubernetes.io/projected/a4422151-89f6-4138-a792-13e1c7473621-kube-api-access-ll2jd\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.593758 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.593768 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4422151-89f6-4138-a792-13e1c7473621-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.595934 4669 scope.go:117] "RemoveContainer" containerID="2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f" Oct 10 09:30:51 crc kubenswrapper[4669]: E1010 09:30:51.596380 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f\": container with ID starting with 2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f not found: ID does not exist" containerID="2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.596407 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f"} err="failed to get container status \"2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f\": rpc error: code = NotFound desc = could not find container \"2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f\": container with ID starting with 2c216013d48075a8c3657473850c54f5c04de12cb6f38c331c0e0eb1c22a7a8f not found: ID does not exist" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.854018 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.861501 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.887042 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:51 crc kubenswrapper[4669]: E1010 09:30:51.921095 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4422151-89f6-4138-a792-13e1c7473621" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.921130 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4422151-89f6-4138-a792-13e1c7473621" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.924393 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4422151-89f6-4138-a792-13e1c7473621" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.925768 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.929145 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.929435 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.929727 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:51 crc kubenswrapper[4669]: I1010 09:30:51.930226 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:51.999955 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.000080 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.000109 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.000176 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdb7b\" (UniqueName: \"kubernetes.io/projected/5d5e59af-66c7-4723-9758-9369ccf46cde-kube-api-access-vdb7b\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.000257 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.101419 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.101820 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.101843 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.101891 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdb7b\" (UniqueName: \"kubernetes.io/projected/5d5e59af-66c7-4723-9758-9369ccf46cde-kube-api-access-vdb7b\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.101943 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.105391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.105454 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.106013 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.115358 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5e59af-66c7-4723-9758-9369ccf46cde-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.123903 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdb7b\" (UniqueName: \"kubernetes.io/projected/5d5e59af-66c7-4723-9758-9369ccf46cde-kube-api-access-vdb7b\") pod \"nova-cell1-novncproxy-0\" (UID: \"5d5e59af-66c7-4723-9758-9369ccf46cde\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.258991 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:52 crc kubenswrapper[4669]: I1010 09:30:52.745092 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.545895 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5d5e59af-66c7-4723-9758-9369ccf46cde","Type":"ContainerStarted","Data":"65eead3073b5c2ab4a993e0d451ff3ccf574b54685bb5969690ec2195209ddd4"} Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.546275 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5d5e59af-66c7-4723-9758-9369ccf46cde","Type":"ContainerStarted","Data":"1b64fd922cbba9530af5a8849842d25b3331a0ff29f7603f5a9527b2e7fbf9cc"} Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.565989 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5659741719999998 podStartE2EDuration="2.565974172s" podCreationTimestamp="2025-10-10 09:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:53.562394907 +0000 UTC m=+1196.578413649" watchObservedRunningTime="2025-10-10 09:30:53.565974172 +0000 UTC m=+1196.581992914" Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.813207 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4422151-89f6-4138-a792-13e1c7473621" path="/var/lib/kubelet/pods/a4422151-89f6-4138-a792-13e1c7473621/volumes" Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.818262 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.819179 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.822070 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 09:30:53 crc kubenswrapper[4669]: I1010 09:30:53.904149 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.555802 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.562836 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.830274 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869677f947-stnwk"] Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.832964 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.849237 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869677f947-stnwk"] Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.948940 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-sb\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.949349 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-nb\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.949396 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rfch\" (UniqueName: \"kubernetes.io/projected/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-kube-api-access-5rfch\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.949564 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-dns-svc\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:54 crc kubenswrapper[4669]: I1010 09:30:54.949698 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-config\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.051402 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-dns-svc\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.051486 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-config\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.051531 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-sb\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.051656 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-nb\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.051712 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rfch\" (UniqueName: \"kubernetes.io/projected/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-kube-api-access-5rfch\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.052372 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-config\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.052433 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-dns-svc\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.052558 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-sb\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.052646 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-nb\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.089946 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rfch\" (UniqueName: \"kubernetes.io/projected/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-kube-api-access-5rfch\") pod \"dnsmasq-dns-869677f947-stnwk\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.165935 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:55 crc kubenswrapper[4669]: I1010 09:30:55.766870 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869677f947-stnwk"] Oct 10 09:30:56 crc kubenswrapper[4669]: I1010 09:30:56.572492 4669 generic.go:334] "Generic (PLEG): container finished" podID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerID="ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744" exitCode=0 Oct 10 09:30:56 crc kubenswrapper[4669]: I1010 09:30:56.572574 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-stnwk" event={"ID":"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d","Type":"ContainerDied","Data":"ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744"} Oct 10 09:30:56 crc kubenswrapper[4669]: I1010 09:30:56.572618 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-stnwk" event={"ID":"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d","Type":"ContainerStarted","Data":"f57001197a30ef2797deaeca4356d102d93780e8bdfb01028b0f8d435bc1e336"} Oct 10 09:30:57 crc kubenswrapper[4669]: I1010 09:30:57.259454 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:30:57 crc kubenswrapper[4669]: I1010 09:30:57.582138 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-stnwk" event={"ID":"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d","Type":"ContainerStarted","Data":"3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63"} Oct 10 09:30:57 crc kubenswrapper[4669]: I1010 09:30:57.582368 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:30:57 crc kubenswrapper[4669]: I1010 09:30:57.609976 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869677f947-stnwk" podStartSLOduration=3.609956935 podStartE2EDuration="3.609956935s" podCreationTimestamp="2025-10-10 09:30:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:30:57.603524979 +0000 UTC m=+1200.619543731" watchObservedRunningTime="2025-10-10 09:30:57.609956935 +0000 UTC m=+1200.625975677" Oct 10 09:30:57 crc kubenswrapper[4669]: I1010 09:30:57.917003 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:30:57 crc kubenswrapper[4669]: I1010 09:30:57.918269 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-api" containerID="cri-o://f4f0191739f96568b968ad3f4831179a17641b91096da7498509d357f8a74e27" gracePeriod=30 Oct 10 09:30:57 crc kubenswrapper[4669]: I1010 09:30:57.918192 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-log" containerID="cri-o://4a0c2697855a129a2341d961950275033c11431bede8136d79befbbf4782fde3" gracePeriod=30 Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.354024 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.354558 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-central-agent" containerID="cri-o://753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b" gracePeriod=30 Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.355035 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="sg-core" containerID="cri-o://b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06" gracePeriod=30 Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.355104 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="proxy-httpd" containerID="cri-o://1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70" gracePeriod=30 Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.355102 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-notification-agent" containerID="cri-o://895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336" gracePeriod=30 Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.395176 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.178:3000/\": EOF" Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.590786 4669 generic.go:334] "Generic (PLEG): container finished" podID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerID="4a0c2697855a129a2341d961950275033c11431bede8136d79befbbf4782fde3" exitCode=143 Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.590844 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"782e0488-3fea-40ae-aeff-766abc1abe9b","Type":"ContainerDied","Data":"4a0c2697855a129a2341d961950275033c11431bede8136d79befbbf4782fde3"} Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.592624 4669 generic.go:334] "Generic (PLEG): container finished" podID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerID="b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06" exitCode=2 Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.593438 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerDied","Data":"b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06"} Oct 10 09:30:58 crc kubenswrapper[4669]: I1010 09:30:58.741564 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.178:3000/\": read tcp 10.217.0.2:43530->10.217.0.178:3000: read: connection reset by peer" Oct 10 09:30:59 crc kubenswrapper[4669]: I1010 09:30:59.602665 4669 generic.go:334] "Generic (PLEG): container finished" podID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerID="1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70" exitCode=0 Oct 10 09:30:59 crc kubenswrapper[4669]: I1010 09:30:59.602694 4669 generic.go:334] "Generic (PLEG): container finished" podID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerID="753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b" exitCode=0 Oct 10 09:30:59 crc kubenswrapper[4669]: I1010 09:30:59.602713 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerDied","Data":"1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70"} Oct 10 09:30:59 crc kubenswrapper[4669]: I1010 09:30:59.602737 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerDied","Data":"753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b"} Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.624226 4669 generic.go:334] "Generic (PLEG): container finished" podID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerID="f4f0191739f96568b968ad3f4831179a17641b91096da7498509d357f8a74e27" exitCode=0 Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.624628 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"782e0488-3fea-40ae-aeff-766abc1abe9b","Type":"ContainerDied","Data":"f4f0191739f96568b968ad3f4831179a17641b91096da7498509d357f8a74e27"} Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.867372 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.985276 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-combined-ca-bundle\") pod \"782e0488-3fea-40ae-aeff-766abc1abe9b\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.985411 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-config-data\") pod \"782e0488-3fea-40ae-aeff-766abc1abe9b\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.985462 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/782e0488-3fea-40ae-aeff-766abc1abe9b-logs\") pod \"782e0488-3fea-40ae-aeff-766abc1abe9b\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.985551 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fctg\" (UniqueName: \"kubernetes.io/projected/782e0488-3fea-40ae-aeff-766abc1abe9b-kube-api-access-4fctg\") pod \"782e0488-3fea-40ae-aeff-766abc1abe9b\" (UID: \"782e0488-3fea-40ae-aeff-766abc1abe9b\") " Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.986049 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/782e0488-3fea-40ae-aeff-766abc1abe9b-logs" (OuterVolumeSpecName: "logs") pod "782e0488-3fea-40ae-aeff-766abc1abe9b" (UID: "782e0488-3fea-40ae-aeff-766abc1abe9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:31:01 crc kubenswrapper[4669]: I1010 09:31:01.987466 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/782e0488-3fea-40ae-aeff-766abc1abe9b-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.022670 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782e0488-3fea-40ae-aeff-766abc1abe9b-kube-api-access-4fctg" (OuterVolumeSpecName: "kube-api-access-4fctg") pod "782e0488-3fea-40ae-aeff-766abc1abe9b" (UID: "782e0488-3fea-40ae-aeff-766abc1abe9b"). InnerVolumeSpecName "kube-api-access-4fctg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.029794 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-config-data" (OuterVolumeSpecName: "config-data") pod "782e0488-3fea-40ae-aeff-766abc1abe9b" (UID: "782e0488-3fea-40ae-aeff-766abc1abe9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.037301 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "782e0488-3fea-40ae-aeff-766abc1abe9b" (UID: "782e0488-3fea-40ae-aeff-766abc1abe9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.092114 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.092623 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/782e0488-3fea-40ae-aeff-766abc1abe9b-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.093861 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fctg\" (UniqueName: \"kubernetes.io/projected/782e0488-3fea-40ae-aeff-766abc1abe9b-kube-api-access-4fctg\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.259562 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.279508 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.634859 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"782e0488-3fea-40ae-aeff-766abc1abe9b","Type":"ContainerDied","Data":"d64019728430070b43e27c1eb6b41bf08e39c8d551b8a94d296ba8978e36f658"} Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.634931 4669 scope.go:117] "RemoveContainer" containerID="f4f0191739f96568b968ad3f4831179a17641b91096da7498509d357f8a74e27" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.634872 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.655663 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.673062 4669 scope.go:117] "RemoveContainer" containerID="4a0c2697855a129a2341d961950275033c11431bede8136d79befbbf4782fde3" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.679863 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.694329 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.714924 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:02 crc kubenswrapper[4669]: E1010 09:31:02.715528 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-api" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.715548 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-api" Oct 10 09:31:02 crc kubenswrapper[4669]: E1010 09:31:02.715592 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-log" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.715600 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-log" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.715819 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-log" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.715837 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" containerName="nova-api-api" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.720061 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.725480 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.732243 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.740474 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.740748 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.826310 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-config-data\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.826418 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-public-tls-certs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.826450 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.826479 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bec59b22-9403-4515-bd11-fcc15b5e594c-logs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.826499 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.826526 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8d55\" (UniqueName: \"kubernetes.io/projected/bec59b22-9403-4515-bd11-fcc15b5e594c-kube-api-access-s8d55\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.906897 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-t6zc6"] Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.908020 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.910097 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.913921 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.927729 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-config-data\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.927807 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-public-tls-certs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.927838 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.927863 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bec59b22-9403-4515-bd11-fcc15b5e594c-logs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.927885 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.927911 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8d55\" (UniqueName: \"kubernetes.io/projected/bec59b22-9403-4515-bd11-fcc15b5e594c-kube-api-access-s8d55\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.929396 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bec59b22-9403-4515-bd11-fcc15b5e594c-logs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.932860 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.933047 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-config-data\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.933257 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.953563 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-public-tls-certs\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.960223 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8d55\" (UniqueName: \"kubernetes.io/projected/bec59b22-9403-4515-bd11-fcc15b5e594c-kube-api-access-s8d55\") pod \"nova-api-0\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " pod="openstack/nova-api-0" Oct 10 09:31:02 crc kubenswrapper[4669]: I1010 09:31:02.963680 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t6zc6"] Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.029822 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.029914 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rclp\" (UniqueName: \"kubernetes.io/projected/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-kube-api-access-9rclp\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.029948 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-config-data\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.029979 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-scripts\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.052159 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.132088 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.132232 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rclp\" (UniqueName: \"kubernetes.io/projected/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-kube-api-access-9rclp\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.132277 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-config-data\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.132338 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-scripts\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.138684 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-config-data\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.139053 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-scripts\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.140307 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.154510 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rclp\" (UniqueName: \"kubernetes.io/projected/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-kube-api-access-9rclp\") pod \"nova-cell1-cell-mapping-t6zc6\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.158934 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.225467 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.335721 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-config-data\") pod \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.335774 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-scripts\") pod \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.335827 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-log-httpd\") pod \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.335939 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-run-httpd\") pod \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.335966 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98q64\" (UniqueName: \"kubernetes.io/projected/a0e4e855-9b34-47f0-9e4b-2e21777e7389-kube-api-access-98q64\") pod \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.336018 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-combined-ca-bundle\") pod \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.336070 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-sg-core-conf-yaml\") pod \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\" (UID: \"a0e4e855-9b34-47f0-9e4b-2e21777e7389\") " Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.337246 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a0e4e855-9b34-47f0-9e4b-2e21777e7389" (UID: "a0e4e855-9b34-47f0-9e4b-2e21777e7389"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.337662 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.341272 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-scripts" (OuterVolumeSpecName: "scripts") pod "a0e4e855-9b34-47f0-9e4b-2e21777e7389" (UID: "a0e4e855-9b34-47f0-9e4b-2e21777e7389"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.341815 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a0e4e855-9b34-47f0-9e4b-2e21777e7389" (UID: "a0e4e855-9b34-47f0-9e4b-2e21777e7389"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.342983 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0e4e855-9b34-47f0-9e4b-2e21777e7389-kube-api-access-98q64" (OuterVolumeSpecName: "kube-api-access-98q64") pod "a0e4e855-9b34-47f0-9e4b-2e21777e7389" (UID: "a0e4e855-9b34-47f0-9e4b-2e21777e7389"). InnerVolumeSpecName "kube-api-access-98q64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.378172 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a0e4e855-9b34-47f0-9e4b-2e21777e7389" (UID: "a0e4e855-9b34-47f0-9e4b-2e21777e7389"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.439640 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.439663 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e4e855-9b34-47f0-9e4b-2e21777e7389-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.439671 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98q64\" (UniqueName: \"kubernetes.io/projected/a0e4e855-9b34-47f0-9e4b-2e21777e7389-kube-api-access-98q64\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.439681 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.476747 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-config-data" (OuterVolumeSpecName: "config-data") pod "a0e4e855-9b34-47f0-9e4b-2e21777e7389" (UID: "a0e4e855-9b34-47f0-9e4b-2e21777e7389"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.476802 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0e4e855-9b34-47f0-9e4b-2e21777e7389" (UID: "a0e4e855-9b34-47f0-9e4b-2e21777e7389"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.499561 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.545456 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.545485 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e4e855-9b34-47f0-9e4b-2e21777e7389-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.648996 4669 generic.go:334] "Generic (PLEG): container finished" podID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerID="895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336" exitCode=0 Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.649382 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerDied","Data":"895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336"} Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.649415 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e4e855-9b34-47f0-9e4b-2e21777e7389","Type":"ContainerDied","Data":"0fea08f3a08cac29659346bc6791217a4351832771bdb12b92b2c676bd122ff1"} Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.649435 4669 scope.go:117] "RemoveContainer" containerID="1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.649532 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.667909 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bec59b22-9403-4515-bd11-fcc15b5e594c","Type":"ContainerStarted","Data":"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6"} Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.667947 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bec59b22-9403-4515-bd11-fcc15b5e594c","Type":"ContainerStarted","Data":"c873adf03d3bd040ad99f692b3d4fe71fbe332265f31483b73d56514f6e2fc1b"} Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.669977 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t6zc6"] Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.702272 4669 scope.go:117] "RemoveContainer" containerID="b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.734807 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.745715 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.753174 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.753732 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="proxy-httpd" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.753753 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="proxy-httpd" Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.753771 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-notification-agent" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.753779 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-notification-agent" Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.753788 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-central-agent" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.753794 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-central-agent" Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.753809 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="sg-core" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.753814 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="sg-core" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.753980 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-central-agent" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.754005 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="sg-core" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.754014 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="proxy-httpd" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.754027 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" containerName="ceilometer-notification-agent" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.755692 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.759833 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.762006 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.762007 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.816238 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782e0488-3fea-40ae-aeff-766abc1abe9b" path="/var/lib/kubelet/pods/782e0488-3fea-40ae-aeff-766abc1abe9b/volumes" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.816977 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0e4e855-9b34-47f0-9e4b-2e21777e7389" path="/var/lib/kubelet/pods/a0e4e855-9b34-47f0-9e4b-2e21777e7389/volumes" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.850935 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.850987 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-config-data\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.851040 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzvqs\" (UniqueName: \"kubernetes.io/projected/465451d1-86d6-47b5-a324-569241f80e33-kube-api-access-lzvqs\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.851196 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-run-httpd\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.851234 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-scripts\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.851294 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.851448 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-log-httpd\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.877185 4669 scope.go:117] "RemoveContainer" containerID="895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.916954 4669 scope.go:117] "RemoveContainer" containerID="753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.940196 4669 scope.go:117] "RemoveContainer" containerID="1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70" Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.940854 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70\": container with ID starting with 1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70 not found: ID does not exist" containerID="1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.940940 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70"} err="failed to get container status \"1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70\": rpc error: code = NotFound desc = could not find container \"1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70\": container with ID starting with 1f42a81f0543a16056e98f4dbadad4fde1d2f6655c7e1775a9abcf5486043b70 not found: ID does not exist" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.940998 4669 scope.go:117] "RemoveContainer" containerID="b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06" Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.941477 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06\": container with ID starting with b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06 not found: ID does not exist" containerID="b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.941508 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06"} err="failed to get container status \"b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06\": rpc error: code = NotFound desc = could not find container \"b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06\": container with ID starting with b9535f38077b00889b68f0199857383472b558012cd8c416cf7f08d596302f06 not found: ID does not exist" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.941530 4669 scope.go:117] "RemoveContainer" containerID="895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336" Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.942042 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336\": container with ID starting with 895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336 not found: ID does not exist" containerID="895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.942068 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336"} err="failed to get container status \"895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336\": rpc error: code = NotFound desc = could not find container \"895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336\": container with ID starting with 895d55d11557f74b44566dd103c888a2637dea1a52ed3e3ef2f3b7b257164336 not found: ID does not exist" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.942081 4669 scope.go:117] "RemoveContainer" containerID="753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b" Oct 10 09:31:03 crc kubenswrapper[4669]: E1010 09:31:03.942329 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b\": container with ID starting with 753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b not found: ID does not exist" containerID="753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.942366 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b"} err="failed to get container status \"753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b\": rpc error: code = NotFound desc = could not find container \"753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b\": container with ID starting with 753e7b42989cc0b295395d75ba71f8ee68e9223aa1f4146a1bdc98bc136c3a5b not found: ID does not exist" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953079 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-run-httpd\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953143 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-scripts\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953200 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953378 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-log-httpd\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953418 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953477 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-config-data\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953500 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzvqs\" (UniqueName: \"kubernetes.io/projected/465451d1-86d6-47b5-a324-569241f80e33-kube-api-access-lzvqs\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.953637 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-run-httpd\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.955199 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-log-httpd\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.959693 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-scripts\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.961278 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.963200 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.968123 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-config-data\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:03 crc kubenswrapper[4669]: I1010 09:31:03.970779 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzvqs\" (UniqueName: \"kubernetes.io/projected/465451d1-86d6-47b5-a324-569241f80e33-kube-api-access-lzvqs\") pod \"ceilometer-0\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " pod="openstack/ceilometer-0" Oct 10 09:31:04 crc kubenswrapper[4669]: I1010 09:31:04.198320 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:04 crc kubenswrapper[4669]: W1010 09:31:04.670638 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod465451d1_86d6_47b5_a324_569241f80e33.slice/crio-bc6f14690796a27f3b5ac8dbcd77d5b09bab1d0bb0d972264d18d9c101df07f1 WatchSource:0}: Error finding container bc6f14690796a27f3b5ac8dbcd77d5b09bab1d0bb0d972264d18d9c101df07f1: Status 404 returned error can't find the container with id bc6f14690796a27f3b5ac8dbcd77d5b09bab1d0bb0d972264d18d9c101df07f1 Oct 10 09:31:04 crc kubenswrapper[4669]: I1010 09:31:04.670921 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:04 crc kubenswrapper[4669]: I1010 09:31:04.679537 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bec59b22-9403-4515-bd11-fcc15b5e594c","Type":"ContainerStarted","Data":"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859"} Oct 10 09:31:04 crc kubenswrapper[4669]: I1010 09:31:04.685844 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t6zc6" event={"ID":"7fc4bf0c-125d-4c30-9389-22b7b80b56f2","Type":"ContainerStarted","Data":"fa3013e1d7a3d6f01698c1fbabf43c0d5477ab4c837c93f6568e4505fa6dd153"} Oct 10 09:31:04 crc kubenswrapper[4669]: I1010 09:31:04.685893 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t6zc6" event={"ID":"7fc4bf0c-125d-4c30-9389-22b7b80b56f2","Type":"ContainerStarted","Data":"734014e36db9db5bcb00cdf1b0b4cfc77f08aa42d42797757e07980de068005d"} Oct 10 09:31:04 crc kubenswrapper[4669]: I1010 09:31:04.700254 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.700233044 podStartE2EDuration="2.700233044s" podCreationTimestamp="2025-10-10 09:31:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:31:04.697161386 +0000 UTC m=+1207.713180118" watchObservedRunningTime="2025-10-10 09:31:04.700233044 +0000 UTC m=+1207.716251786" Oct 10 09:31:04 crc kubenswrapper[4669]: I1010 09:31:04.721649 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-t6zc6" podStartSLOduration=2.721631432 podStartE2EDuration="2.721631432s" podCreationTimestamp="2025-10-10 09:31:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:31:04.719713121 +0000 UTC m=+1207.735731863" watchObservedRunningTime="2025-10-10 09:31:04.721631432 +0000 UTC m=+1207.737650174" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.167783 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.230046 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-mbghh"] Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.230322 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" podUID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerName="dnsmasq-dns" containerID="cri-o://f8ed530963ea3feec5e9535df8e231e53a2312c374c2402469b81a4dea85de07" gracePeriod=10 Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.713061 4669 generic.go:334] "Generic (PLEG): container finished" podID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerID="f8ed530963ea3feec5e9535df8e231e53a2312c374c2402469b81a4dea85de07" exitCode=0 Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.713259 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" event={"ID":"aab7b93a-d05b-484b-b93a-ada52b09b9d2","Type":"ContainerDied","Data":"f8ed530963ea3feec5e9535df8e231e53a2312c374c2402469b81a4dea85de07"} Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.715522 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerStarted","Data":"bc6f14690796a27f3b5ac8dbcd77d5b09bab1d0bb0d972264d18d9c101df07f1"} Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.829062 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.889040 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-dns-svc\") pod \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.889095 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-nb\") pod \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.889297 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-sb\") pod \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.889335 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-config\") pod \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.889400 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrrq8\" (UniqueName: \"kubernetes.io/projected/aab7b93a-d05b-484b-b93a-ada52b09b9d2-kube-api-access-jrrq8\") pod \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\" (UID: \"aab7b93a-d05b-484b-b93a-ada52b09b9d2\") " Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.906520 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab7b93a-d05b-484b-b93a-ada52b09b9d2-kube-api-access-jrrq8" (OuterVolumeSpecName: "kube-api-access-jrrq8") pod "aab7b93a-d05b-484b-b93a-ada52b09b9d2" (UID: "aab7b93a-d05b-484b-b93a-ada52b09b9d2"). InnerVolumeSpecName "kube-api-access-jrrq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.959858 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aab7b93a-d05b-484b-b93a-ada52b09b9d2" (UID: "aab7b93a-d05b-484b-b93a-ada52b09b9d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.972351 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aab7b93a-d05b-484b-b93a-ada52b09b9d2" (UID: "aab7b93a-d05b-484b-b93a-ada52b09b9d2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.980546 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aab7b93a-d05b-484b-b93a-ada52b09b9d2" (UID: "aab7b93a-d05b-484b-b93a-ada52b09b9d2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.987222 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-config" (OuterVolumeSpecName: "config") pod "aab7b93a-d05b-484b-b93a-ada52b09b9d2" (UID: "aab7b93a-d05b-484b-b93a-ada52b09b9d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.991835 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.991868 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.991879 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrrq8\" (UniqueName: \"kubernetes.io/projected/aab7b93a-d05b-484b-b93a-ada52b09b9d2-kube-api-access-jrrq8\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.991890 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:05 crc kubenswrapper[4669]: I1010 09:31:05.991899 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aab7b93a-d05b-484b-b93a-ada52b09b9d2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.724124 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" event={"ID":"aab7b93a-d05b-484b-b93a-ada52b09b9d2","Type":"ContainerDied","Data":"b8fe88eff24a6817a2dad61bc76926577413d866573b25c64c86fe603db9fe28"} Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.724535 4669 scope.go:117] "RemoveContainer" containerID="f8ed530963ea3feec5e9535df8e231e53a2312c374c2402469b81a4dea85de07" Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.724136 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-mbghh" Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.725886 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerStarted","Data":"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da"} Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.725914 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerStarted","Data":"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528"} Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.741124 4669 scope.go:117] "RemoveContainer" containerID="38cd74d82a03a716dbf4a22a88b9523dbf74a938b974d3b831a0ba8bfbdda5c6" Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.765799 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-mbghh"] Oct 10 09:31:06 crc kubenswrapper[4669]: I1010 09:31:06.786572 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-mbghh"] Oct 10 09:31:07 crc kubenswrapper[4669]: I1010 09:31:07.739425 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerStarted","Data":"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af"} Oct 10 09:31:07 crc kubenswrapper[4669]: I1010 09:31:07.846156 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" path="/var/lib/kubelet/pods/aab7b93a-d05b-484b-b93a-ada52b09b9d2/volumes" Oct 10 09:31:09 crc kubenswrapper[4669]: I1010 09:31:09.767118 4669 generic.go:334] "Generic (PLEG): container finished" podID="7fc4bf0c-125d-4c30-9389-22b7b80b56f2" containerID="fa3013e1d7a3d6f01698c1fbabf43c0d5477ab4c837c93f6568e4505fa6dd153" exitCode=0 Oct 10 09:31:09 crc kubenswrapper[4669]: I1010 09:31:09.768293 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t6zc6" event={"ID":"7fc4bf0c-125d-4c30-9389-22b7b80b56f2","Type":"ContainerDied","Data":"fa3013e1d7a3d6f01698c1fbabf43c0d5477ab4c837c93f6568e4505fa6dd153"} Oct 10 09:31:09 crc kubenswrapper[4669]: I1010 09:31:09.771878 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerStarted","Data":"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f"} Oct 10 09:31:09 crc kubenswrapper[4669]: I1010 09:31:09.773128 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 09:31:09 crc kubenswrapper[4669]: I1010 09:31:09.830939 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3984275520000002 podStartE2EDuration="6.830913907s" podCreationTimestamp="2025-10-10 09:31:03 +0000 UTC" firstStartedPulling="2025-10-10 09:31:04.677558785 +0000 UTC m=+1207.693577527" lastFinishedPulling="2025-10-10 09:31:09.11004514 +0000 UTC m=+1212.126063882" observedRunningTime="2025-10-10 09:31:09.819318334 +0000 UTC m=+1212.835337076" watchObservedRunningTime="2025-10-10 09:31:09.830913907 +0000 UTC m=+1212.846932649" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.104409 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.188710 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-combined-ca-bundle\") pod \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.189632 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rclp\" (UniqueName: \"kubernetes.io/projected/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-kube-api-access-9rclp\") pod \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.189704 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-scripts\") pod \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.189846 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-config-data\") pod \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\" (UID: \"7fc4bf0c-125d-4c30-9389-22b7b80b56f2\") " Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.194447 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-scripts" (OuterVolumeSpecName: "scripts") pod "7fc4bf0c-125d-4c30-9389-22b7b80b56f2" (UID: "7fc4bf0c-125d-4c30-9389-22b7b80b56f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.201291 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-kube-api-access-9rclp" (OuterVolumeSpecName: "kube-api-access-9rclp") pod "7fc4bf0c-125d-4c30-9389-22b7b80b56f2" (UID: "7fc4bf0c-125d-4c30-9389-22b7b80b56f2"). InnerVolumeSpecName "kube-api-access-9rclp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.227787 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fc4bf0c-125d-4c30-9389-22b7b80b56f2" (UID: "7fc4bf0c-125d-4c30-9389-22b7b80b56f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.244363 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-config-data" (OuterVolumeSpecName: "config-data") pod "7fc4bf0c-125d-4c30-9389-22b7b80b56f2" (UID: "7fc4bf0c-125d-4c30-9389-22b7b80b56f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.292336 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.292379 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.292395 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rclp\" (UniqueName: \"kubernetes.io/projected/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-kube-api-access-9rclp\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.292406 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc4bf0c-125d-4c30-9389-22b7b80b56f2-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.793243 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t6zc6" event={"ID":"7fc4bf0c-125d-4c30-9389-22b7b80b56f2","Type":"ContainerDied","Data":"734014e36db9db5bcb00cdf1b0b4cfc77f08aa42d42797757e07980de068005d"} Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.793599 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="734014e36db9db5bcb00cdf1b0b4cfc77f08aa42d42797757e07980de068005d" Oct 10 09:31:11 crc kubenswrapper[4669]: I1010 09:31:11.793667 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t6zc6" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.015779 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.016291 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-log" containerID="cri-o://f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6" gracePeriod=30 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.016361 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-api" containerID="cri-o://3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859" gracePeriod=30 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.036216 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.036479 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" containerName="nova-scheduler-scheduler" containerID="cri-o://8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" gracePeriod=30 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.072781 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.073010 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-log" containerID="cri-o://202bf293457faa36e046bce932af6e750f2bbb5c297dfe57f8d63ffc21b6e0be" gracePeriod=30 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.073138 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-metadata" containerID="cri-o://cf98206075bede22a040c201cc3d0a2111f3715fc14c2ba83e06b94d116d4a43" gracePeriod=30 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.676057 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.832745 4669 generic.go:334] "Generic (PLEG): container finished" podID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerID="202bf293457faa36e046bce932af6e750f2bbb5c297dfe57f8d63ffc21b6e0be" exitCode=143 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.833093 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93f92d98-29f6-44b2-9087-1dbd9d74aa62","Type":"ContainerDied","Data":"202bf293457faa36e046bce932af6e750f2bbb5c297dfe57f8d63ffc21b6e0be"} Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.837219 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-public-tls-certs\") pod \"bec59b22-9403-4515-bd11-fcc15b5e594c\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.837293 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-internal-tls-certs\") pod \"bec59b22-9403-4515-bd11-fcc15b5e594c\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.837340 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-config-data\") pod \"bec59b22-9403-4515-bd11-fcc15b5e594c\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.837370 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8d55\" (UniqueName: \"kubernetes.io/projected/bec59b22-9403-4515-bd11-fcc15b5e594c-kube-api-access-s8d55\") pod \"bec59b22-9403-4515-bd11-fcc15b5e594c\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.837444 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-combined-ca-bundle\") pod \"bec59b22-9403-4515-bd11-fcc15b5e594c\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.837465 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bec59b22-9403-4515-bd11-fcc15b5e594c-logs\") pod \"bec59b22-9403-4515-bd11-fcc15b5e594c\" (UID: \"bec59b22-9403-4515-bd11-fcc15b5e594c\") " Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.838180 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bec59b22-9403-4515-bd11-fcc15b5e594c-logs" (OuterVolumeSpecName: "logs") pod "bec59b22-9403-4515-bd11-fcc15b5e594c" (UID: "bec59b22-9403-4515-bd11-fcc15b5e594c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.843858 4669 generic.go:334] "Generic (PLEG): container finished" podID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerID="3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859" exitCode=0 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.843896 4669 generic.go:334] "Generic (PLEG): container finished" podID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerID="f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6" exitCode=143 Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.843916 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bec59b22-9403-4515-bd11-fcc15b5e594c","Type":"ContainerDied","Data":"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859"} Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.843941 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bec59b22-9403-4515-bd11-fcc15b5e594c","Type":"ContainerDied","Data":"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6"} Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.843952 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bec59b22-9403-4515-bd11-fcc15b5e594c","Type":"ContainerDied","Data":"c873adf03d3bd040ad99f692b3d4fe71fbe332265f31483b73d56514f6e2fc1b"} Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.843966 4669 scope.go:117] "RemoveContainer" containerID="3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.844169 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.858154 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bec59b22-9403-4515-bd11-fcc15b5e594c-kube-api-access-s8d55" (OuterVolumeSpecName: "kube-api-access-s8d55") pod "bec59b22-9403-4515-bd11-fcc15b5e594c" (UID: "bec59b22-9403-4515-bd11-fcc15b5e594c"). InnerVolumeSpecName "kube-api-access-s8d55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.935211 4669 scope.go:117] "RemoveContainer" containerID="f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.939458 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8d55\" (UniqueName: \"kubernetes.io/projected/bec59b22-9403-4515-bd11-fcc15b5e594c-kube-api-access-s8d55\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.939484 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bec59b22-9403-4515-bd11-fcc15b5e594c-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.949400 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bec59b22-9403-4515-bd11-fcc15b5e594c" (UID: "bec59b22-9403-4515-bd11-fcc15b5e594c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.954829 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bec59b22-9403-4515-bd11-fcc15b5e594c" (UID: "bec59b22-9403-4515-bd11-fcc15b5e594c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.956332 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-config-data" (OuterVolumeSpecName: "config-data") pod "bec59b22-9403-4515-bd11-fcc15b5e594c" (UID: "bec59b22-9403-4515-bd11-fcc15b5e594c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.974961 4669 scope.go:117] "RemoveContainer" containerID="3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859" Oct 10 09:31:12 crc kubenswrapper[4669]: E1010 09:31:12.976733 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859\": container with ID starting with 3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859 not found: ID does not exist" containerID="3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.976770 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859"} err="failed to get container status \"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859\": rpc error: code = NotFound desc = could not find container \"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859\": container with ID starting with 3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859 not found: ID does not exist" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.976793 4669 scope.go:117] "RemoveContainer" containerID="f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6" Oct 10 09:31:12 crc kubenswrapper[4669]: E1010 09:31:12.978906 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6\": container with ID starting with f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6 not found: ID does not exist" containerID="f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.978943 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6"} err="failed to get container status \"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6\": rpc error: code = NotFound desc = could not find container \"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6\": container with ID starting with f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6 not found: ID does not exist" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.978963 4669 scope.go:117] "RemoveContainer" containerID="3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.981901 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859"} err="failed to get container status \"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859\": rpc error: code = NotFound desc = could not find container \"3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859\": container with ID starting with 3734b9623c32780e7b6e82c71797b778fcb5d09a0d3e05a73215ac4a1f866859 not found: ID does not exist" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.981937 4669 scope.go:117] "RemoveContainer" containerID="f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6" Oct 10 09:31:12 crc kubenswrapper[4669]: I1010 09:31:12.982275 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6"} err="failed to get container status \"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6\": rpc error: code = NotFound desc = could not find container \"f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6\": container with ID starting with f4982a326bd46ace4208abd962c1f0f7e3ef01a03ee42cf6668162a96cf592b6 not found: ID does not exist" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.004683 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bec59b22-9403-4515-bd11-fcc15b5e594c" (UID: "bec59b22-9403-4515-bd11-fcc15b5e594c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.041542 4669 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.041613 4669 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.041623 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.041631 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec59b22-9403-4515-bd11-fcc15b5e594c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.197892 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.204154 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219024 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.219384 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-log" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219400 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-log" Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.219412 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc4bf0c-125d-4c30-9389-22b7b80b56f2" containerName="nova-manage" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219420 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc4bf0c-125d-4c30-9389-22b7b80b56f2" containerName="nova-manage" Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.219431 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-api" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219439 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-api" Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.219453 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerName="init" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219459 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerName="init" Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.219476 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerName="dnsmasq-dns" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219482 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerName="dnsmasq-dns" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219649 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc4bf0c-125d-4c30-9389-22b7b80b56f2" containerName="nova-manage" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219666 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-log" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219674 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" containerName="nova-api-api" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.219686 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab7b93a-d05b-484b-b93a-ada52b09b9d2" containerName="dnsmasq-dns" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.220531 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.224892 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.225163 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.238486 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.254866 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.355015 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw75l\" (UniqueName: \"kubernetes.io/projected/d81e619a-c272-4824-b19f-dea88a428acd-kube-api-access-lw75l\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.355108 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.355179 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.355232 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-public-tls-certs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.355259 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d81e619a-c272-4824-b19f-dea88a428acd-logs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.355284 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-config-data\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.457672 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.457766 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.457819 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-public-tls-certs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.457848 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d81e619a-c272-4824-b19f-dea88a428acd-logs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.457873 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-config-data\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.457947 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw75l\" (UniqueName: \"kubernetes.io/projected/d81e619a-c272-4824-b19f-dea88a428acd-kube-api-access-lw75l\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.458368 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d81e619a-c272-4824-b19f-dea88a428acd-logs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.466477 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-config-data\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.466568 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.479806 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-public-tls-certs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.480531 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d81e619a-c272-4824-b19f-dea88a428acd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.483335 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw75l\" (UniqueName: \"kubernetes.io/projected/d81e619a-c272-4824-b19f-dea88a428acd-kube-api-access-lw75l\") pod \"nova-api-0\" (UID: \"d81e619a-c272-4824-b19f-dea88a428acd\") " pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.571565 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.769255 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254 is running failed: container process not found" containerID="8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.769917 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254 is running failed: container process not found" containerID="8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.770343 4669 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254 is running failed: container process not found" containerID="8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.770375 4669 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" containerName="nova-scheduler-scheduler" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.783730 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.822437 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bec59b22-9403-4515-bd11-fcc15b5e594c" path="/var/lib/kubelet/pods/bec59b22-9403-4515-bd11-fcc15b5e594c/volumes" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.856113 4669 generic.go:334] "Generic (PLEG): container finished" podID="2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" containerID="8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" exitCode=0 Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.856177 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e","Type":"ContainerDied","Data":"8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254"} Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.856205 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e","Type":"ContainerDied","Data":"e4c1d22b2a5e002eb55abb1ccae5e28fd9a1f556b6d51cac93ed98deba04507b"} Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.856220 4669 scope.go:117] "RemoveContainer" containerID="8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.856300 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.879021 4669 scope.go:117] "RemoveContainer" containerID="8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" Oct 10 09:31:13 crc kubenswrapper[4669]: E1010 09:31:13.879476 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254\": container with ID starting with 8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254 not found: ID does not exist" containerID="8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.879522 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254"} err="failed to get container status \"8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254\": rpc error: code = NotFound desc = could not find container \"8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254\": container with ID starting with 8cc590fde6791c06bbfd18736b8cf972b9293d3e3eecada541fddc3ec282d254 not found: ID does not exist" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.967101 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-config-data\") pod \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.967331 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f67wt\" (UniqueName: \"kubernetes.io/projected/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-kube-api-access-f67wt\") pod \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.967406 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-combined-ca-bundle\") pod \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\" (UID: \"2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e\") " Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.974542 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-kube-api-access-f67wt" (OuterVolumeSpecName: "kube-api-access-f67wt") pod "2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" (UID: "2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e"). InnerVolumeSpecName "kube-api-access-f67wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.994727 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-config-data" (OuterVolumeSpecName: "config-data") pod "2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" (UID: "2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:13 crc kubenswrapper[4669]: I1010 09:31:13.998601 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" (UID: "2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.069271 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.069308 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.069322 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f67wt\" (UniqueName: \"kubernetes.io/projected/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e-kube-api-access-f67wt\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:14 crc kubenswrapper[4669]: W1010 09:31:14.116044 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd81e619a_c272_4824_b19f_dea88a428acd.slice/crio-de5aa7893e9fc8d8107fd7d38e2a00524535e986f5cad8d56f4df1db429547ba WatchSource:0}: Error finding container de5aa7893e9fc8d8107fd7d38e2a00524535e986f5cad8d56f4df1db429547ba: Status 404 returned error can't find the container with id de5aa7893e9fc8d8107fd7d38e2a00524535e986f5cad8d56f4df1db429547ba Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.118004 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.258548 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.275209 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.292968 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:31:14 crc kubenswrapper[4669]: E1010 09:31:14.293329 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" containerName="nova-scheduler-scheduler" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.293344 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" containerName="nova-scheduler-scheduler" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.293513 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" containerName="nova-scheduler-scheduler" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.294078 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.297062 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.340428 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.477641 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rwz4\" (UniqueName: \"kubernetes.io/projected/cbfa50a5-b829-4023-b978-d76660fc77b2-kube-api-access-9rwz4\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.477880 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfa50a5-b829-4023-b978-d76660fc77b2-config-data\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.477933 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfa50a5-b829-4023-b978-d76660fc77b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.579609 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rwz4\" (UniqueName: \"kubernetes.io/projected/cbfa50a5-b829-4023-b978-d76660fc77b2-kube-api-access-9rwz4\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.579677 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfa50a5-b829-4023-b978-d76660fc77b2-config-data\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.579732 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfa50a5-b829-4023-b978-d76660fc77b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.583813 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfa50a5-b829-4023-b978-d76660fc77b2-config-data\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.588228 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbfa50a5-b829-4023-b978-d76660fc77b2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.596671 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rwz4\" (UniqueName: \"kubernetes.io/projected/cbfa50a5-b829-4023-b978-d76660fc77b2-kube-api-access-9rwz4\") pod \"nova-scheduler-0\" (UID: \"cbfa50a5-b829-4023-b978-d76660fc77b2\") " pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.610059 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.877429 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d81e619a-c272-4824-b19f-dea88a428acd","Type":"ContainerStarted","Data":"bee8793613749bd721cb4de38e8f33788ae359f414315f9ceef553db31c1ce03"} Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.877792 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d81e619a-c272-4824-b19f-dea88a428acd","Type":"ContainerStarted","Data":"d9f50417f1e3f0fdbb1fb600090955b5977910f2d253c583c03d2642f05b373a"} Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.877809 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d81e619a-c272-4824-b19f-dea88a428acd","Type":"ContainerStarted","Data":"de5aa7893e9fc8d8107fd7d38e2a00524535e986f5cad8d56f4df1db429547ba"} Oct 10 09:31:14 crc kubenswrapper[4669]: I1010 09:31:14.901639 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.901618532 podStartE2EDuration="1.901618532s" podCreationTimestamp="2025-10-10 09:31:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:31:14.896936341 +0000 UTC m=+1217.912955073" watchObservedRunningTime="2025-10-10 09:31:14.901618532 +0000 UTC m=+1217.917637274" Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.088381 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 09:31:15 crc kubenswrapper[4669]: W1010 09:31:15.095977 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbfa50a5_b829_4023_b978_d76660fc77b2.slice/crio-f43ed1bfc8e0f1e17241823cda0e0b67b3642d17b0719354041d12c5cb031ac5 WatchSource:0}: Error finding container f43ed1bfc8e0f1e17241823cda0e0b67b3642d17b0719354041d12c5cb031ac5: Status 404 returned error can't find the container with id f43ed1bfc8e0f1e17241823cda0e0b67b3642d17b0719354041d12c5cb031ac5 Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.808641 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e" path="/var/lib/kubelet/pods/2c34e6e6-4020-40a8-ae2c-9bf3e7ddcf2e/volumes" Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.891633 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbfa50a5-b829-4023-b978-d76660fc77b2","Type":"ContainerStarted","Data":"52b6a0a0843e8709733f8b27181be77d695260b718621621f91872f4959de7e1"} Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.891670 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbfa50a5-b829-4023-b978-d76660fc77b2","Type":"ContainerStarted","Data":"f43ed1bfc8e0f1e17241823cda0e0b67b3642d17b0719354041d12c5cb031ac5"} Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.905282 4669 generic.go:334] "Generic (PLEG): container finished" podID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerID="cf98206075bede22a040c201cc3d0a2111f3715fc14c2ba83e06b94d116d4a43" exitCode=0 Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.905944 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93f92d98-29f6-44b2-9087-1dbd9d74aa62","Type":"ContainerDied","Data":"cf98206075bede22a040c201cc3d0a2111f3715fc14c2ba83e06b94d116d4a43"} Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.905973 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"93f92d98-29f6-44b2-9087-1dbd9d74aa62","Type":"ContainerDied","Data":"82137489e3dd19ea7f930324656e7c39f29afaa23ad813e997e404c1fbce341d"} Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.905985 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82137489e3dd19ea7f930324656e7c39f29afaa23ad813e997e404c1fbce341d" Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.912038 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.912025429 podStartE2EDuration="1.912025429s" podCreationTimestamp="2025-10-10 09:31:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:31:15.910875222 +0000 UTC m=+1218.926893964" watchObservedRunningTime="2025-10-10 09:31:15.912025429 +0000 UTC m=+1218.928044161" Oct 10 09:31:15 crc kubenswrapper[4669]: I1010 09:31:15.938825 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.106159 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwjrb\" (UniqueName: \"kubernetes.io/projected/93f92d98-29f6-44b2-9087-1dbd9d74aa62-kube-api-access-kwjrb\") pod \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.106451 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-nova-metadata-tls-certs\") pod \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.106510 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-config-data\") pod \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.106531 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-combined-ca-bundle\") pod \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.106596 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93f92d98-29f6-44b2-9087-1dbd9d74aa62-logs\") pod \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\" (UID: \"93f92d98-29f6-44b2-9087-1dbd9d74aa62\") " Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.107290 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93f92d98-29f6-44b2-9087-1dbd9d74aa62-logs" (OuterVolumeSpecName: "logs") pod "93f92d98-29f6-44b2-9087-1dbd9d74aa62" (UID: "93f92d98-29f6-44b2-9087-1dbd9d74aa62"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.110859 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f92d98-29f6-44b2-9087-1dbd9d74aa62-kube-api-access-kwjrb" (OuterVolumeSpecName: "kube-api-access-kwjrb") pod "93f92d98-29f6-44b2-9087-1dbd9d74aa62" (UID: "93f92d98-29f6-44b2-9087-1dbd9d74aa62"). InnerVolumeSpecName "kube-api-access-kwjrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.133502 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-config-data" (OuterVolumeSpecName: "config-data") pod "93f92d98-29f6-44b2-9087-1dbd9d74aa62" (UID: "93f92d98-29f6-44b2-9087-1dbd9d74aa62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.145669 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93f92d98-29f6-44b2-9087-1dbd9d74aa62" (UID: "93f92d98-29f6-44b2-9087-1dbd9d74aa62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.178026 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "93f92d98-29f6-44b2-9087-1dbd9d74aa62" (UID: "93f92d98-29f6-44b2-9087-1dbd9d74aa62"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.209623 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwjrb\" (UniqueName: \"kubernetes.io/projected/93f92d98-29f6-44b2-9087-1dbd9d74aa62-kube-api-access-kwjrb\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.209651 4669 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.209660 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.209671 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93f92d98-29f6-44b2-9087-1dbd9d74aa62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.209679 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93f92d98-29f6-44b2-9087-1dbd9d74aa62-logs\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.915681 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.951079 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.963729 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.986893 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:31:16 crc kubenswrapper[4669]: E1010 09:31:16.987395 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-metadata" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.987421 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-metadata" Oct 10 09:31:16 crc kubenswrapper[4669]: E1010 09:31:16.987459 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-log" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.987470 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-log" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.990173 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-metadata" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.990234 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-log" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.992348 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.996850 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 09:31:16 crc kubenswrapper[4669]: I1010 09:31:16.999750 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.013339 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.124539 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2jlm\" (UniqueName: \"kubernetes.io/projected/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-kube-api-access-d2jlm\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.124609 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.124682 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-logs\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.124720 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-config-data\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.124773 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.226654 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2jlm\" (UniqueName: \"kubernetes.io/projected/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-kube-api-access-d2jlm\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.226998 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.227114 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-logs\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.227238 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-config-data\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.227376 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.229113 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-logs\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.233985 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.234157 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-config-data\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.242687 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.245860 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2jlm\" (UniqueName: \"kubernetes.io/projected/f6b1ea56-8a03-4b04-ba07-0852d3fb6545-kube-api-access-d2jlm\") pod \"nova-metadata-0\" (UID: \"f6b1ea56-8a03-4b04-ba07-0852d3fb6545\") " pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.321080 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.814615 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" path="/var/lib/kubelet/pods/93f92d98-29f6-44b2-9087-1dbd9d74aa62/volumes" Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.815342 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 09:31:17 crc kubenswrapper[4669]: I1010 09:31:17.927954 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b1ea56-8a03-4b04-ba07-0852d3fb6545","Type":"ContainerStarted","Data":"478f0a1437a1850bc342b335643d2223caff8a330473a171a67cf8c6a65f9f54"} Oct 10 09:31:18 crc kubenswrapper[4669]: I1010 09:31:18.940694 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b1ea56-8a03-4b04-ba07-0852d3fb6545","Type":"ContainerStarted","Data":"2485e113329ba97de0df733615f3b4aa5d0cdce32b65f4f975006b57c27598b9"} Oct 10 09:31:18 crc kubenswrapper[4669]: I1010 09:31:18.942196 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b1ea56-8a03-4b04-ba07-0852d3fb6545","Type":"ContainerStarted","Data":"a864b6690d9a6ec04cbfb1ed6f173dae8c15f006f8d827afa3105e9493b53794"} Oct 10 09:31:18 crc kubenswrapper[4669]: I1010 09:31:18.974632 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.974567887 podStartE2EDuration="2.974567887s" podCreationTimestamp="2025-10-10 09:31:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:31:18.961905569 +0000 UTC m=+1221.977924361" watchObservedRunningTime="2025-10-10 09:31:18.974567887 +0000 UTC m=+1221.990586659" Oct 10 09:31:19 crc kubenswrapper[4669]: I1010 09:31:19.610906 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 09:31:20 crc kubenswrapper[4669]: I1010 09:31:20.877580 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 09:31:20 crc kubenswrapper[4669]: I1010 09:31:20.877638 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="93f92d98-29f6-44b2-9087-1dbd9d74aa62" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 09:31:22 crc kubenswrapper[4669]: I1010 09:31:22.322210 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 09:31:22 crc kubenswrapper[4669]: I1010 09:31:22.324347 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 09:31:23 crc kubenswrapper[4669]: I1010 09:31:23.572753 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 09:31:23 crc kubenswrapper[4669]: I1010 09:31:23.572828 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 09:31:24 crc kubenswrapper[4669]: I1010 09:31:24.582759 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d81e619a-c272-4824-b19f-dea88a428acd" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 09:31:24 crc kubenswrapper[4669]: I1010 09:31:24.582760 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d81e619a-c272-4824-b19f-dea88a428acd" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.188:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 09:31:24 crc kubenswrapper[4669]: I1010 09:31:24.610253 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 09:31:24 crc kubenswrapper[4669]: I1010 09:31:24.636219 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 09:31:25 crc kubenswrapper[4669]: I1010 09:31:25.022093 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 09:31:27 crc kubenswrapper[4669]: I1010 09:31:27.321802 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 09:31:27 crc kubenswrapper[4669]: I1010 09:31:27.322152 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 09:31:28 crc kubenswrapper[4669]: I1010 09:31:28.359260 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f6b1ea56-8a03-4b04-ba07-0852d3fb6545" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 09:31:28 crc kubenswrapper[4669]: I1010 09:31:28.359291 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f6b1ea56-8a03-4b04-ba07-0852d3fb6545" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 09:31:33 crc kubenswrapper[4669]: I1010 09:31:33.584811 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 09:31:33 crc kubenswrapper[4669]: I1010 09:31:33.585781 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 09:31:33 crc kubenswrapper[4669]: I1010 09:31:33.587335 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 09:31:33 crc kubenswrapper[4669]: I1010 09:31:33.594968 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 09:31:34 crc kubenswrapper[4669]: I1010 09:31:34.096311 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 09:31:34 crc kubenswrapper[4669]: I1010 09:31:34.108894 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 09:31:34 crc kubenswrapper[4669]: I1010 09:31:34.218441 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.031176 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.119447 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" containerName="kube-state-metrics" containerID="cri-o://1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34" gracePeriod=30 Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.329005 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.364008 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.365742 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.627118 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.745973 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l9k8\" (UniqueName: \"kubernetes.io/projected/e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79-kube-api-access-2l9k8\") pod \"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79\" (UID: \"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79\") " Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.752912 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79-kube-api-access-2l9k8" (OuterVolumeSpecName: "kube-api-access-2l9k8") pod "e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" (UID: "e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79"). InnerVolumeSpecName "kube-api-access-2l9k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:37 crc kubenswrapper[4669]: I1010 09:31:37.848168 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l9k8\" (UniqueName: \"kubernetes.io/projected/e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79-kube-api-access-2l9k8\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.129524 4669 generic.go:334] "Generic (PLEG): container finished" podID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" containerID="1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34" exitCode=2 Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.129602 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.129677 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79","Type":"ContainerDied","Data":"1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34"} Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.129741 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79","Type":"ContainerDied","Data":"84363c2a9cb53453a784ec043bfe54489c4a63d2070cbf73761f8c2bbf68f56f"} Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.129759 4669 scope.go:117] "RemoveContainer" containerID="1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.131147 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.131394 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-central-agent" containerID="cri-o://e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" gracePeriod=30 Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.131430 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-notification-agent" containerID="cri-o://9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" gracePeriod=30 Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.131404 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="proxy-httpd" containerID="cri-o://97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" gracePeriod=30 Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.131435 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="sg-core" containerID="cri-o://0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" gracePeriod=30 Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.144841 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.156868 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.166713 4669 scope.go:117] "RemoveContainer" containerID="1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34" Oct 10 09:31:38 crc kubenswrapper[4669]: E1010 09:31:38.167173 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34\": container with ID starting with 1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34 not found: ID does not exist" containerID="1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.167215 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34"} err="failed to get container status \"1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34\": rpc error: code = NotFound desc = could not find container \"1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34\": container with ID starting with 1a8598b46160e10e3991706c338b3eecbd21cebcde4ba6969524870c6c579f34 not found: ID does not exist" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.174124 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.182962 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:31:38 crc kubenswrapper[4669]: E1010 09:31:38.183409 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" containerName="kube-state-metrics" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.183429 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" containerName="kube-state-metrics" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.183663 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" containerName="kube-state-metrics" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.184287 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.189320 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.189393 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.270261 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.359388 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.360083 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.360300 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.360452 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwm52\" (UniqueName: \"kubernetes.io/projected/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-api-access-bwm52\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.461945 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwm52\" (UniqueName: \"kubernetes.io/projected/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-api-access-bwm52\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.462681 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.466494 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.466661 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.471511 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.473135 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.481001 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwm52\" (UniqueName: \"kubernetes.io/projected/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-api-access-bwm52\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.487934 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0330c42a-ef18-46cd-9a95-5622b53328b7-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0330c42a-ef18-46cd-9a95-5622b53328b7\") " pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.508629 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 09:31:38 crc kubenswrapper[4669]: I1010 09:31:38.996164 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.072862 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139478 4669 generic.go:334] "Generic (PLEG): container finished" podID="465451d1-86d6-47b5-a324-569241f80e33" containerID="97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" exitCode=0 Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139508 4669 generic.go:334] "Generic (PLEG): container finished" podID="465451d1-86d6-47b5-a324-569241f80e33" containerID="0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" exitCode=2 Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139514 4669 generic.go:334] "Generic (PLEG): container finished" podID="465451d1-86d6-47b5-a324-569241f80e33" containerID="9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" exitCode=0 Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139523 4669 generic.go:334] "Generic (PLEG): container finished" podID="465451d1-86d6-47b5-a324-569241f80e33" containerID="e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" exitCode=0 Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139560 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerDied","Data":"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f"} Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139599 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerDied","Data":"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af"} Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139609 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerDied","Data":"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da"} Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139619 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerDied","Data":"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528"} Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139628 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"465451d1-86d6-47b5-a324-569241f80e33","Type":"ContainerDied","Data":"bc6f14690796a27f3b5ac8dbcd77d5b09bab1d0bb0d972264d18d9c101df07f1"} Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139643 4669 scope.go:117] "RemoveContainer" containerID="97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.139766 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.144870 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0330c42a-ef18-46cd-9a95-5622b53328b7","Type":"ContainerStarted","Data":"1c3134b7029cab4271aa7ce243ed2e18c3610084abae38ecccd0f76345f26d87"} Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.158724 4669 scope.go:117] "RemoveContainer" containerID="0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.177771 4669 scope.go:117] "RemoveContainer" containerID="9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.184241 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-sg-core-conf-yaml\") pod \"465451d1-86d6-47b5-a324-569241f80e33\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.184407 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-config-data\") pod \"465451d1-86d6-47b5-a324-569241f80e33\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.184485 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzvqs\" (UniqueName: \"kubernetes.io/projected/465451d1-86d6-47b5-a324-569241f80e33-kube-api-access-lzvqs\") pod \"465451d1-86d6-47b5-a324-569241f80e33\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.184515 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-combined-ca-bundle\") pod \"465451d1-86d6-47b5-a324-569241f80e33\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.184594 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-run-httpd\") pod \"465451d1-86d6-47b5-a324-569241f80e33\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.185384 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-scripts\") pod \"465451d1-86d6-47b5-a324-569241f80e33\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.185409 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-log-httpd\") pod \"465451d1-86d6-47b5-a324-569241f80e33\" (UID: \"465451d1-86d6-47b5-a324-569241f80e33\") " Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.188069 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "465451d1-86d6-47b5-a324-569241f80e33" (UID: "465451d1-86d6-47b5-a324-569241f80e33"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.188258 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "465451d1-86d6-47b5-a324-569241f80e33" (UID: "465451d1-86d6-47b5-a324-569241f80e33"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.191460 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-scripts" (OuterVolumeSpecName: "scripts") pod "465451d1-86d6-47b5-a324-569241f80e33" (UID: "465451d1-86d6-47b5-a324-569241f80e33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.191833 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/465451d1-86d6-47b5-a324-569241f80e33-kube-api-access-lzvqs" (OuterVolumeSpecName: "kube-api-access-lzvqs") pod "465451d1-86d6-47b5-a324-569241f80e33" (UID: "465451d1-86d6-47b5-a324-569241f80e33"). InnerVolumeSpecName "kube-api-access-lzvqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.204155 4669 scope.go:117] "RemoveContainer" containerID="e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.225573 4669 scope.go:117] "RemoveContainer" containerID="97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.226103 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": container with ID starting with 97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f not found: ID does not exist" containerID="97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.227390 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f"} err="failed to get container status \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": rpc error: code = NotFound desc = could not find container \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": container with ID starting with 97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.227495 4669 scope.go:117] "RemoveContainer" containerID="0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.227895 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": container with ID starting with 0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af not found: ID does not exist" containerID="0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.227916 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af"} err="failed to get container status \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": rpc error: code = NotFound desc = could not find container \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": container with ID starting with 0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.227936 4669 scope.go:117] "RemoveContainer" containerID="9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.228137 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": container with ID starting with 9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da not found: ID does not exist" containerID="9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.228157 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da"} err="failed to get container status \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": rpc error: code = NotFound desc = could not find container \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": container with ID starting with 9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.228170 4669 scope.go:117] "RemoveContainer" containerID="e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.228403 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": container with ID starting with e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528 not found: ID does not exist" containerID="e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.228513 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528"} err="failed to get container status \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": rpc error: code = NotFound desc = could not find container \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": container with ID starting with e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528 not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.231708 4669 scope.go:117] "RemoveContainer" containerID="97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.232010 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f"} err="failed to get container status \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": rpc error: code = NotFound desc = could not find container \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": container with ID starting with 97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.232102 4669 scope.go:117] "RemoveContainer" containerID="0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.232386 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af"} err="failed to get container status \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": rpc error: code = NotFound desc = could not find container \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": container with ID starting with 0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.232424 4669 scope.go:117] "RemoveContainer" containerID="9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.232635 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da"} err="failed to get container status \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": rpc error: code = NotFound desc = could not find container \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": container with ID starting with 9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.232747 4669 scope.go:117] "RemoveContainer" containerID="e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.232998 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528"} err="failed to get container status \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": rpc error: code = NotFound desc = could not find container \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": container with ID starting with e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528 not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.233089 4669 scope.go:117] "RemoveContainer" containerID="97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.233494 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f"} err="failed to get container status \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": rpc error: code = NotFound desc = could not find container \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": container with ID starting with 97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.233615 4669 scope.go:117] "RemoveContainer" containerID="0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.233999 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af"} err="failed to get container status \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": rpc error: code = NotFound desc = could not find container \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": container with ID starting with 0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.234019 4669 scope.go:117] "RemoveContainer" containerID="9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.234298 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da"} err="failed to get container status \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": rpc error: code = NotFound desc = could not find container \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": container with ID starting with 9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.234387 4669 scope.go:117] "RemoveContainer" containerID="e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.234691 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528"} err="failed to get container status \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": rpc error: code = NotFound desc = could not find container \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": container with ID starting with e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528 not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.234730 4669 scope.go:117] "RemoveContainer" containerID="97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.234959 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f"} err="failed to get container status \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": rpc error: code = NotFound desc = could not find container \"97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f\": container with ID starting with 97b0e85c7ea7f654a09ab5565e33a537899d3ad6fc560cc4cc6e038623df707f not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.235073 4669 scope.go:117] "RemoveContainer" containerID="0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.235333 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af"} err="failed to get container status \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": rpc error: code = NotFound desc = could not find container \"0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af\": container with ID starting with 0bc3c2d9f009d03ae659b41d7b31ce0ef8364988d85e25128b98f94051bd92af not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.235366 4669 scope.go:117] "RemoveContainer" containerID="9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.235658 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da"} err="failed to get container status \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": rpc error: code = NotFound desc = could not find container \"9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da\": container with ID starting with 9baa88404ed0245f20c85f1abebd558be103ff3531df6233e984c10c31c0e8da not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.235800 4669 scope.go:117] "RemoveContainer" containerID="e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.236143 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528"} err="failed to get container status \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": rpc error: code = NotFound desc = could not find container \"e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528\": container with ID starting with e5480d12bab59668e1af8e3517d3474c9f6844364e5fd050b1869a5bc1da9528 not found: ID does not exist" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.243288 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "465451d1-86d6-47b5-a324-569241f80e33" (UID: "465451d1-86d6-47b5-a324-569241f80e33"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.288081 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "465451d1-86d6-47b5-a324-569241f80e33" (UID: "465451d1-86d6-47b5-a324-569241f80e33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.288393 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzvqs\" (UniqueName: \"kubernetes.io/projected/465451d1-86d6-47b5-a324-569241f80e33-kube-api-access-lzvqs\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.288439 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.288449 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.288459 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465451d1-86d6-47b5-a324-569241f80e33-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.288466 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.323464 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-config-data" (OuterVolumeSpecName: "config-data") pod "465451d1-86d6-47b5-a324-569241f80e33" (UID: "465451d1-86d6-47b5-a324-569241f80e33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.390563 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.390884 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465451d1-86d6-47b5-a324-569241f80e33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.531330 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.540322 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.559402 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.559750 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="sg-core" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.559766 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="sg-core" Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.559776 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-notification-agent" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.559782 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-notification-agent" Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.559790 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-central-agent" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.559797 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-central-agent" Oct 10 09:31:39 crc kubenswrapper[4669]: E1010 09:31:39.559833 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="proxy-httpd" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.559839 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="proxy-httpd" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.559987 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="proxy-httpd" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.560006 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-notification-agent" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.560015 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="sg-core" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.560026 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="465451d1-86d6-47b5-a324-569241f80e33" containerName="ceilometer-central-agent" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.561526 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.563861 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.564122 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.564974 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.583005 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593550 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-config-data\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593658 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593687 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-run-httpd\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593721 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593771 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-log-httpd\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593816 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-scripts\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593878 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.593912 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8txcp\" (UniqueName: \"kubernetes.io/projected/386e6beb-5cc7-46d0-a201-008319df489e-kube-api-access-8txcp\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695403 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-log-httpd\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695481 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-scripts\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695541 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695577 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8txcp\" (UniqueName: \"kubernetes.io/projected/386e6beb-5cc7-46d0-a201-008319df489e-kube-api-access-8txcp\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695649 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-config-data\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695685 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695707 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-run-httpd\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.695734 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.696896 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-run-httpd\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.697154 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-log-httpd\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.701679 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.701844 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.701901 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.703321 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-scripts\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.711920 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-config-data\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.714935 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8txcp\" (UniqueName: \"kubernetes.io/projected/386e6beb-5cc7-46d0-a201-008319df489e-kube-api-access-8txcp\") pod \"ceilometer-0\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " pod="openstack/ceilometer-0" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.806564 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="465451d1-86d6-47b5-a324-569241f80e33" path="/var/lib/kubelet/pods/465451d1-86d6-47b5-a324-569241f80e33/volumes" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.808449 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79" path="/var/lib/kubelet/pods/e8508ed5-1c27-4ce7-b5da-abfbb3bb0b79/volumes" Oct 10 09:31:39 crc kubenswrapper[4669]: I1010 09:31:39.883239 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 09:31:40 crc kubenswrapper[4669]: I1010 09:31:40.163530 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0330c42a-ef18-46cd-9a95-5622b53328b7","Type":"ContainerStarted","Data":"29ac3d365fd280743583513b1913725ec524758aa2b1b928002792f70665041b"} Oct 10 09:31:40 crc kubenswrapper[4669]: I1010 09:31:40.163829 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 10 09:31:40 crc kubenswrapper[4669]: I1010 09:31:40.185230 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.781570012 podStartE2EDuration="2.185211279s" podCreationTimestamp="2025-10-10 09:31:38 +0000 UTC" firstStartedPulling="2025-10-10 09:31:39.006436558 +0000 UTC m=+1242.022455300" lastFinishedPulling="2025-10-10 09:31:39.410077835 +0000 UTC m=+1242.426096567" observedRunningTime="2025-10-10 09:31:40.184233447 +0000 UTC m=+1243.200252199" watchObservedRunningTime="2025-10-10 09:31:40.185211279 +0000 UTC m=+1243.201230021" Oct 10 09:31:40 crc kubenswrapper[4669]: I1010 09:31:40.328833 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 09:31:41 crc kubenswrapper[4669]: I1010 09:31:41.028244 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:31:41 crc kubenswrapper[4669]: I1010 09:31:41.171288 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerStarted","Data":"b40d7bcd632155f9ffca95723093002b995fb03d2736c6b048a1f27e594e39be"} Oct 10 09:31:41 crc kubenswrapper[4669]: I1010 09:31:41.171337 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerStarted","Data":"3a9553be6248c003a7a80166b2ff2d2ae4cd52ac5e30e5905c507a7f3dc041fa"} Oct 10 09:31:42 crc kubenswrapper[4669]: I1010 09:31:42.182082 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerStarted","Data":"7f9bd1b093c85231fbebb5d1df480ec463cd334b04a08c2f0550744d2a6f14e8"} Oct 10 09:31:43 crc kubenswrapper[4669]: I1010 09:31:43.192706 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerStarted","Data":"710ea080bd15ec03e3840b34e1a2004e345429d935400bdc67e3f5c4e9ff941c"} Oct 10 09:31:44 crc kubenswrapper[4669]: I1010 09:31:44.205136 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerStarted","Data":"54bfbc40fe297f1b48b18ac6cdffd691ed80efa7f722956a5ba5106dac538192"} Oct 10 09:31:44 crc kubenswrapper[4669]: I1010 09:31:44.205471 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 09:31:44 crc kubenswrapper[4669]: I1010 09:31:44.259406 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8804030900000002 podStartE2EDuration="5.259380352s" podCreationTimestamp="2025-10-10 09:31:39 +0000 UTC" firstStartedPulling="2025-10-10 09:31:40.338111984 +0000 UTC m=+1243.354130726" lastFinishedPulling="2025-10-10 09:31:43.717089246 +0000 UTC m=+1246.733107988" observedRunningTime="2025-10-10 09:31:44.251407306 +0000 UTC m=+1247.267426058" watchObservedRunningTime="2025-10-10 09:31:44.259380352 +0000 UTC m=+1247.275399104" Oct 10 09:31:48 crc kubenswrapper[4669]: I1010 09:31:48.519479 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 10 09:32:09 crc kubenswrapper[4669]: I1010 09:32:09.892368 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 09:32:18 crc kubenswrapper[4669]: I1010 09:32:18.696660 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:32:19 crc kubenswrapper[4669]: I1010 09:32:19.643920 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:32:22 crc kubenswrapper[4669]: I1010 09:32:22.924544 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerName="rabbitmq" containerID="cri-o://e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80" gracePeriod=604796 Oct 10 09:32:23 crc kubenswrapper[4669]: I1010 09:32:23.032851 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 10 09:32:24 crc kubenswrapper[4669]: I1010 09:32:24.031349 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerName="rabbitmq" containerID="cri-o://3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc" gracePeriod=604796 Oct 10 09:32:24 crc kubenswrapper[4669]: I1010 09:32:24.275103 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:32:24 crc kubenswrapper[4669]: I1010 09:32:24.275554 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.465938 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.638506 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-plugins-conf\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.638895 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b787ed5f-cbcf-4144-a87f-a014dbfd5282-erlang-cookie-secret\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.638940 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2m77\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-kube-api-access-j2m77\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.638958 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-config-data\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.638992 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b787ed5f-cbcf-4144-a87f-a014dbfd5282-pod-info\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.639052 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-plugins\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.639079 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-erlang-cookie\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.639098 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-tls\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.639128 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.639148 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-server-conf\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.639182 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-confd\") pod \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\" (UID: \"b787ed5f-cbcf-4144-a87f-a014dbfd5282\") " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.641120 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.643214 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.645830 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.645998 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b787ed5f-cbcf-4144-a87f-a014dbfd5282-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.647042 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.647169 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b787ed5f-cbcf-4144-a87f-a014dbfd5282-pod-info" (OuterVolumeSpecName: "pod-info") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.647953 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-kube-api-access-j2m77" (OuterVolumeSpecName: "kube-api-access-j2m77") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "kube-api-access-j2m77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.655183 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.658882 4669 generic.go:334] "Generic (PLEG): container finished" podID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerID="e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80" exitCode=0 Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.658918 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b787ed5f-cbcf-4144-a87f-a014dbfd5282","Type":"ContainerDied","Data":"e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80"} Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.658943 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b787ed5f-cbcf-4144-a87f-a014dbfd5282","Type":"ContainerDied","Data":"5309828c5f2cfa424e41988a4bf0ee40352f6d3e358be206e07bb604e0a1e1a0"} Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.658942 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.658956 4669 scope.go:117] "RemoveContainer" containerID="e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.677994 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-config-data" (OuterVolumeSpecName: "config-data") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.704097 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-server-conf" (OuterVolumeSpecName: "server-conf") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741101 4669 scope.go:117] "RemoveContainer" containerID="971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741872 4669 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741900 4669 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b787ed5f-cbcf-4144-a87f-a014dbfd5282-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741911 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2m77\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-kube-api-access-j2m77\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741920 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741929 4669 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b787ed5f-cbcf-4144-a87f-a014dbfd5282-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741937 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741945 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741953 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741977 4669 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.741985 4669 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b787ed5f-cbcf-4144-a87f-a014dbfd5282-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.761431 4669 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.761531 4669 scope.go:117] "RemoveContainer" containerID="e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80" Oct 10 09:32:29 crc kubenswrapper[4669]: E1010 09:32:29.763833 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80\": container with ID starting with e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80 not found: ID does not exist" containerID="e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.763883 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80"} err="failed to get container status \"e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80\": rpc error: code = NotFound desc = could not find container \"e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80\": container with ID starting with e6f88ba4ab469bb51db3655c0257074a0c205a0831c2088faa850e5deb4d0b80 not found: ID does not exist" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.763919 4669 scope.go:117] "RemoveContainer" containerID="971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3" Oct 10 09:32:29 crc kubenswrapper[4669]: E1010 09:32:29.765483 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3\": container with ID starting with 971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3 not found: ID does not exist" containerID="971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.765505 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3"} err="failed to get container status \"971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3\": rpc error: code = NotFound desc = could not find container \"971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3\": container with ID starting with 971a33512293002443f6373ff59cb81d5e6051d44b157d8ba665889b3485a2f3 not found: ID does not exist" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.793740 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b787ed5f-cbcf-4144-a87f-a014dbfd5282" (UID: "b787ed5f-cbcf-4144-a87f-a014dbfd5282"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.844034 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b787ed5f-cbcf-4144-a87f-a014dbfd5282-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.844276 4669 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.981457 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:32:29 crc kubenswrapper[4669]: I1010 09:32:29.988162 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.010285 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:32:30 crc kubenswrapper[4669]: E1010 09:32:30.010763 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerName="rabbitmq" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.010781 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerName="rabbitmq" Oct 10 09:32:30 crc kubenswrapper[4669]: E1010 09:32:30.010816 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerName="setup-container" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.010825 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerName="setup-container" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.011061 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" containerName="rabbitmq" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.012453 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.015100 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.015484 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.015620 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.015720 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.015837 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.015947 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.016092 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-drd74" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.026995 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158242 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158312 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158414 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158507 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158613 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1172c8c-d2a5-4f42-b97a-87265385580b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158768 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158803 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7599m\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-kube-api-access-7599m\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158850 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.158937 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-config-data\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.159019 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.159087 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1172c8c-d2a5-4f42-b97a-87265385580b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260397 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260772 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7599m\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-kube-api-access-7599m\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260799 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260857 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-config-data\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260900 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260919 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1172c8c-d2a5-4f42-b97a-87265385580b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260936 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260960 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.260990 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.261009 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.261027 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1172c8c-d2a5-4f42-b97a-87265385580b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.261074 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.261331 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.261579 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.261712 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.262746 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.263416 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c1172c8c-d2a5-4f42-b97a-87265385580b-config-data\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.266839 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c1172c8c-d2a5-4f42-b97a-87265385580b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.267016 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c1172c8c-d2a5-4f42-b97a-87265385580b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.270926 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.271428 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.283905 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7599m\" (UniqueName: \"kubernetes.io/projected/c1172c8c-d2a5-4f42-b97a-87265385580b-kube-api-access-7599m\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.302015 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"c1172c8c-d2a5-4f42-b97a-87265385580b\") " pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.337310 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.545046 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.678940 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-erlang-cookie\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679037 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c46f84af-548a-4390-8242-c48ff4fc25aa-erlang-cookie-secret\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679101 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-config-data\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679152 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2rhr\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-kube-api-access-c2rhr\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679201 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-confd\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679260 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-tls\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679321 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-plugins\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679389 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-server-conf\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679432 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-plugins-conf\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679511 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.679545 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c46f84af-548a-4390-8242-c48ff4fc25aa-pod-info\") pod \"c46f84af-548a-4390-8242-c48ff4fc25aa\" (UID: \"c46f84af-548a-4390-8242-c48ff4fc25aa\") " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.686734 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.687349 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.688003 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.690317 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.691260 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c46f84af-548a-4390-8242-c48ff4fc25aa-pod-info" (OuterVolumeSpecName: "pod-info") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.691541 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.697810 4669 generic.go:334] "Generic (PLEG): container finished" podID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerID="3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc" exitCode=0 Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.698103 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.698686 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-kube-api-access-c2rhr" (OuterVolumeSpecName: "kube-api-access-c2rhr") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "kube-api-access-c2rhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.698771 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c46f84af-548a-4390-8242-c48ff4fc25aa","Type":"ContainerDied","Data":"3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc"} Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.698814 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c46f84af-548a-4390-8242-c48ff4fc25aa","Type":"ContainerDied","Data":"8844181a1c7977635a20bf8cbb536885fe0b9c3fe8d3afc8b91e19cbfb50fc07"} Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.698837 4669 scope.go:117] "RemoveContainer" containerID="3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.707401 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c46f84af-548a-4390-8242-c48ff4fc25aa-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.740808 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-config-data" (OuterVolumeSpecName: "config-data") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.747936 4669 scope.go:117] "RemoveContainer" containerID="c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.769787 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-server-conf" (OuterVolumeSpecName: "server-conf") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781302 4669 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781337 4669 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c46f84af-548a-4390-8242-c48ff4fc25aa-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781350 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781361 4669 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c46f84af-548a-4390-8242-c48ff4fc25aa-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781370 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781379 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2rhr\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-kube-api-access-c2rhr\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781387 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781395 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781403 4669 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.781411 4669 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c46f84af-548a-4390-8242-c48ff4fc25aa-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.784032 4669 scope.go:117] "RemoveContainer" containerID="3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc" Oct 10 09:32:30 crc kubenswrapper[4669]: E1010 09:32:30.784510 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc\": container with ID starting with 3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc not found: ID does not exist" containerID="3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.784552 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc"} err="failed to get container status \"3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc\": rpc error: code = NotFound desc = could not find container \"3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc\": container with ID starting with 3eb337ce75564b8a35891e26d47c8cd12b18ee3850f39ce8cd0d2e693f4d51cc not found: ID does not exist" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.790720 4669 scope.go:117] "RemoveContainer" containerID="c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8" Oct 10 09:32:30 crc kubenswrapper[4669]: E1010 09:32:30.796254 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8\": container with ID starting with c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8 not found: ID does not exist" containerID="c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.796302 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8"} err="failed to get container status \"c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8\": rpc error: code = NotFound desc = could not find container \"c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8\": container with ID starting with c3fb9c331c3a939f3dff122bc1d5a51e831ad824f2a015fb03f131f53ed89ac8 not found: ID does not exist" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.805405 4669 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.829653 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c46f84af-548a-4390-8242-c48ff4fc25aa" (UID: "c46f84af-548a-4390-8242-c48ff4fc25aa"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.882718 4669 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.882753 4669 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c46f84af-548a-4390-8242-c48ff4fc25aa-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:30 crc kubenswrapper[4669]: I1010 09:32:30.896359 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.065790 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.082936 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.090873 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:32:31 crc kubenswrapper[4669]: E1010 09:32:31.091224 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerName="setup-container" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.091243 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerName="setup-container" Oct 10 09:32:31 crc kubenswrapper[4669]: E1010 09:32:31.091255 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerName="rabbitmq" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.091261 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerName="rabbitmq" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.091415 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="c46f84af-548a-4390-8242-c48ff4fc25aa" containerName="rabbitmq" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.093798 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.098337 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.098443 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.098551 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.098746 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.098862 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.098919 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.100164 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-9rdrl" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.110123 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186611 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186673 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05416508-674b-44d8-9074-62776915ec92-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186704 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186737 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05416508-674b-44d8-9074-62776915ec92-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186792 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186808 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05416508-674b-44d8-9074-62776915ec92-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186852 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05416508-674b-44d8-9074-62776915ec92-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186883 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186907 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.186982 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.187030 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9svv\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-kube-api-access-m9svv\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288736 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288817 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288845 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288868 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9svv\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-kube-api-access-m9svv\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288898 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288914 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05416508-674b-44d8-9074-62776915ec92-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288933 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.288964 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05416508-674b-44d8-9074-62776915ec92-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.289016 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.289036 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05416508-674b-44d8-9074-62776915ec92-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.289077 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05416508-674b-44d8-9074-62776915ec92-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.289493 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.289774 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.290994 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05416508-674b-44d8-9074-62776915ec92-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.291310 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.291487 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05416508-674b-44d8-9074-62776915ec92-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.293522 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05416508-674b-44d8-9074-62776915ec92-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.294343 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05416508-674b-44d8-9074-62776915ec92-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.295870 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.299102 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05416508-674b-44d8-9074-62776915ec92-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.299819 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.308846 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9svv\" (UniqueName: \"kubernetes.io/projected/05416508-674b-44d8-9074-62776915ec92-kube-api-access-m9svv\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.318322 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"05416508-674b-44d8-9074-62776915ec92\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.417361 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.710510 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1172c8c-d2a5-4f42-b97a-87265385580b","Type":"ContainerStarted","Data":"4ddc3096a3fe802b93966a13e831e92b28f489d59a3e1229fb585193e1ba2dd1"} Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.807849 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b787ed5f-cbcf-4144-a87f-a014dbfd5282" path="/var/lib/kubelet/pods/b787ed5f-cbcf-4144-a87f-a014dbfd5282/volumes" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.808639 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c46f84af-548a-4390-8242-c48ff4fc25aa" path="/var/lib/kubelet/pods/c46f84af-548a-4390-8242-c48ff4fc25aa/volumes" Oct 10 09:32:31 crc kubenswrapper[4669]: I1010 09:32:31.857402 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 09:32:31 crc kubenswrapper[4669]: W1010 09:32:31.963508 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05416508_674b_44d8_9074_62776915ec92.slice/crio-e85287c51914952cde8ae3c5845082f30b8cec1559e578f224b1596cb7c5b053 WatchSource:0}: Error finding container e85287c51914952cde8ae3c5845082f30b8cec1559e578f224b1596cb7c5b053: Status 404 returned error can't find the container with id e85287c51914952cde8ae3c5845082f30b8cec1559e578f224b1596cb7c5b053 Oct 10 09:32:32 crc kubenswrapper[4669]: I1010 09:32:32.720036 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05416508-674b-44d8-9074-62776915ec92","Type":"ContainerStarted","Data":"e85287c51914952cde8ae3c5845082f30b8cec1559e578f224b1596cb7c5b053"} Oct 10 09:32:32 crc kubenswrapper[4669]: I1010 09:32:32.722339 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1172c8c-d2a5-4f42-b97a-87265385580b","Type":"ContainerStarted","Data":"41748360c429b731cab9fa6e47baa4fdeb44a2b9e5c7c023cfb593301f2a12bf"} Oct 10 09:32:33 crc kubenswrapper[4669]: I1010 09:32:33.734219 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05416508-674b-44d8-9074-62776915ec92","Type":"ContainerStarted","Data":"ba4e4454caac1c3b0dc597fad0f94916957f952155c50a70eafd78f05dc88d0e"} Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.763378 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-pfcv6"] Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.765390 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.770054 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.775991 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-pfcv6"] Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.867249 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9v9s\" (UniqueName: \"kubernetes.io/projected/4ade1efd-98f9-4aa8-a415-6bac982b95d8-kube-api-access-h9v9s\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.867475 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-openstack-edpm-ipam\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.867549 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-config\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.867606 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-sb\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.867685 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-dns-svc\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.867711 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-nb\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.969652 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-openstack-edpm-ipam\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.969995 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-config\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.970146 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-sb\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.970300 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-dns-svc\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.970398 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-nb\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.970536 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9v9s\" (UniqueName: \"kubernetes.io/projected/4ade1efd-98f9-4aa8-a415-6bac982b95d8-kube-api-access-h9v9s\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.970562 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-openstack-edpm-ipam\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.971182 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-dns-svc\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.971291 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-sb\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.971332 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-nb\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:34 crc kubenswrapper[4669]: I1010 09:32:34.971680 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-config\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:35 crc kubenswrapper[4669]: I1010 09:32:35.019655 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9v9s\" (UniqueName: \"kubernetes.io/projected/4ade1efd-98f9-4aa8-a415-6bac982b95d8-kube-api-access-h9v9s\") pod \"dnsmasq-dns-5745cbd8d7-pfcv6\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:35 crc kubenswrapper[4669]: I1010 09:32:35.092665 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:35 crc kubenswrapper[4669]: I1010 09:32:35.584017 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-pfcv6"] Oct 10 09:32:35 crc kubenswrapper[4669]: I1010 09:32:35.756212 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" event={"ID":"4ade1efd-98f9-4aa8-a415-6bac982b95d8","Type":"ContainerStarted","Data":"5fef36905607f1d5580a400444633fc06ffc48286a52391f6464d12cc5809b29"} Oct 10 09:32:36 crc kubenswrapper[4669]: I1010 09:32:36.771101 4669 generic.go:334] "Generic (PLEG): container finished" podID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerID="f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97" exitCode=0 Oct 10 09:32:36 crc kubenswrapper[4669]: I1010 09:32:36.771186 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" event={"ID":"4ade1efd-98f9-4aa8-a415-6bac982b95d8","Type":"ContainerDied","Data":"f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97"} Oct 10 09:32:37 crc kubenswrapper[4669]: I1010 09:32:37.780371 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" event={"ID":"4ade1efd-98f9-4aa8-a415-6bac982b95d8","Type":"ContainerStarted","Data":"c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d"} Oct 10 09:32:37 crc kubenswrapper[4669]: I1010 09:32:37.780849 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:37 crc kubenswrapper[4669]: I1010 09:32:37.813110 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" podStartSLOduration=3.813090093 podStartE2EDuration="3.813090093s" podCreationTimestamp="2025-10-10 09:32:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:32:37.805639614 +0000 UTC m=+1300.821658416" watchObservedRunningTime="2025-10-10 09:32:37.813090093 +0000 UTC m=+1300.829108835" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.095903 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.187169 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869677f947-stnwk"] Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.187401 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869677f947-stnwk" podUID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerName="dnsmasq-dns" containerID="cri-o://3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63" gracePeriod=10 Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.396176 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d4c9c94c9-dbk8f"] Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.403289 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.435212 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d4c9c94c9-dbk8f"] Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.598431 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.598485 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-dns-svc\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.598519 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqzkt\" (UniqueName: \"kubernetes.io/projected/7bc42ad9-e7d1-4579-ab07-f0b625372429-kube-api-access-wqzkt\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.598557 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-nb\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.598737 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-sb\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.598849 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-config\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.700672 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-nb\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.701078 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-sb\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.701680 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-nb\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.701829 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-sb\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.702707 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-config\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.702767 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.702794 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-dns-svc\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.702816 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqzkt\" (UniqueName: \"kubernetes.io/projected/7bc42ad9-e7d1-4579-ab07-f0b625372429-kube-api-access-wqzkt\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.704145 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.704785 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-dns-svc\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.704883 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-config\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.735257 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqzkt\" (UniqueName: \"kubernetes.io/projected/7bc42ad9-e7d1-4579-ab07-f0b625372429-kube-api-access-wqzkt\") pod \"dnsmasq-dns-6d4c9c94c9-dbk8f\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.809315 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.901018 4669 generic.go:334] "Generic (PLEG): container finished" podID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerID="3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63" exitCode=0 Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.901065 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-stnwk" event={"ID":"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d","Type":"ContainerDied","Data":"3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63"} Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.901092 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-stnwk" event={"ID":"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d","Type":"ContainerDied","Data":"f57001197a30ef2797deaeca4356d102d93780e8bdfb01028b0f8d435bc1e336"} Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.901092 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-stnwk" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.901109 4669 scope.go:117] "RemoveContainer" containerID="3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.923552 4669 scope.go:117] "RemoveContainer" containerID="ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.946004 4669 scope.go:117] "RemoveContainer" containerID="3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63" Oct 10 09:32:45 crc kubenswrapper[4669]: E1010 09:32:45.947002 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63\": container with ID starting with 3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63 not found: ID does not exist" containerID="3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.947053 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63"} err="failed to get container status \"3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63\": rpc error: code = NotFound desc = could not find container \"3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63\": container with ID starting with 3825fad4b86f4c55716a47df542203dd4cce39d712819e2325e08c4704dd7e63 not found: ID does not exist" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.947087 4669 scope.go:117] "RemoveContainer" containerID="ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744" Oct 10 09:32:45 crc kubenswrapper[4669]: E1010 09:32:45.947508 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744\": container with ID starting with ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744 not found: ID does not exist" containerID="ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744" Oct 10 09:32:45 crc kubenswrapper[4669]: I1010 09:32:45.947559 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744"} err="failed to get container status \"ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744\": rpc error: code = NotFound desc = could not find container \"ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744\": container with ID starting with ac9a1fd786236f7583c8888bd309e30a91eaf5f84c17c47382a6125e57fd3744 not found: ID does not exist" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.008534 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rfch\" (UniqueName: \"kubernetes.io/projected/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-kube-api-access-5rfch\") pod \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.008680 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-dns-svc\") pod \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.008773 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-nb\") pod \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.008795 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-sb\") pod \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.008945 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-config\") pod \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\" (UID: \"2b9f4b3b-8d7c-499f-a32f-8662c0ae873d\") " Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.012998 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-kube-api-access-5rfch" (OuterVolumeSpecName: "kube-api-access-5rfch") pod "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" (UID: "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d"). InnerVolumeSpecName "kube-api-access-5rfch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.033060 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.058792 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-config" (OuterVolumeSpecName: "config") pod "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" (UID: "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.063954 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" (UID: "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.067103 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" (UID: "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.070057 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" (UID: "2b9f4b3b-8d7c-499f-a32f-8662c0ae873d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.110664 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.110703 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rfch\" (UniqueName: \"kubernetes.io/projected/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-kube-api-access-5rfch\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.110715 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.110724 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.110732 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.287171 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869677f947-stnwk"] Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.296579 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869677f947-stnwk"] Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.474632 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d4c9c94c9-dbk8f"] Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.914710 4669 generic.go:334] "Generic (PLEG): container finished" podID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerID="5bec3eae24e2177201f6883216457a655f42fea3f002bf2c850ef14e6fbe3541" exitCode=0 Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.914800 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" event={"ID":"7bc42ad9-e7d1-4579-ab07-f0b625372429","Type":"ContainerDied","Data":"5bec3eae24e2177201f6883216457a655f42fea3f002bf2c850ef14e6fbe3541"} Oct 10 09:32:46 crc kubenswrapper[4669]: I1010 09:32:46.915122 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" event={"ID":"7bc42ad9-e7d1-4579-ab07-f0b625372429","Type":"ContainerStarted","Data":"fea1e8ee6f8d160215f8ee79a78ee52133558a8a9012e926cc5f4617bb021ac1"} Oct 10 09:32:47 crc kubenswrapper[4669]: I1010 09:32:47.820936 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" path="/var/lib/kubelet/pods/2b9f4b3b-8d7c-499f-a32f-8662c0ae873d/volumes" Oct 10 09:32:47 crc kubenswrapper[4669]: I1010 09:32:47.935103 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" event={"ID":"7bc42ad9-e7d1-4579-ab07-f0b625372429","Type":"ContainerStarted","Data":"7bb695aaeecf2c868512e35ab57bc26f1de3403e4962da54add9988f5584e6be"} Oct 10 09:32:47 crc kubenswrapper[4669]: I1010 09:32:47.935441 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:54 crc kubenswrapper[4669]: I1010 09:32:54.274769 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:32:54 crc kubenswrapper[4669]: I1010 09:32:54.275206 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.034870 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.064977 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" podStartSLOduration=11.064947501 podStartE2EDuration="11.064947501s" podCreationTimestamp="2025-10-10 09:32:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:32:47.961046281 +0000 UTC m=+1310.977065063" watchObservedRunningTime="2025-10-10 09:32:56.064947501 +0000 UTC m=+1319.080966283" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.129430 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-pfcv6"] Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.129711 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" podUID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerName="dnsmasq-dns" containerID="cri-o://c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d" gracePeriod=10 Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.688630 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.751824 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-openstack-edpm-ipam\") pod \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.751898 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-sb\") pod \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.751947 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-dns-svc\") pod \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.751987 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-nb\") pod \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.752062 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9v9s\" (UniqueName: \"kubernetes.io/projected/4ade1efd-98f9-4aa8-a415-6bac982b95d8-kube-api-access-h9v9s\") pod \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.752103 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-config\") pod \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\" (UID: \"4ade1efd-98f9-4aa8-a415-6bac982b95d8\") " Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.763890 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ade1efd-98f9-4aa8-a415-6bac982b95d8-kube-api-access-h9v9s" (OuterVolumeSpecName: "kube-api-access-h9v9s") pod "4ade1efd-98f9-4aa8-a415-6bac982b95d8" (UID: "4ade1efd-98f9-4aa8-a415-6bac982b95d8"). InnerVolumeSpecName "kube-api-access-h9v9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.813970 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "4ade1efd-98f9-4aa8-a415-6bac982b95d8" (UID: "4ade1efd-98f9-4aa8-a415-6bac982b95d8"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.828020 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-config" (OuterVolumeSpecName: "config") pod "4ade1efd-98f9-4aa8-a415-6bac982b95d8" (UID: "4ade1efd-98f9-4aa8-a415-6bac982b95d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.829071 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ade1efd-98f9-4aa8-a415-6bac982b95d8" (UID: "4ade1efd-98f9-4aa8-a415-6bac982b95d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.839211 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ade1efd-98f9-4aa8-a415-6bac982b95d8" (UID: "4ade1efd-98f9-4aa8-a415-6bac982b95d8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.848919 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ade1efd-98f9-4aa8-a415-6bac982b95d8" (UID: "4ade1efd-98f9-4aa8-a415-6bac982b95d8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.854550 4669 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.854594 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.854606 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.854615 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.854625 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9v9s\" (UniqueName: \"kubernetes.io/projected/4ade1efd-98f9-4aa8-a415-6bac982b95d8-kube-api-access-h9v9s\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:56 crc kubenswrapper[4669]: I1010 09:32:56.854635 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ade1efd-98f9-4aa8-a415-6bac982b95d8-config\") on node \"crc\" DevicePath \"\"" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.026696 4669 generic.go:334] "Generic (PLEG): container finished" podID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerID="c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d" exitCode=0 Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.026741 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" event={"ID":"4ade1efd-98f9-4aa8-a415-6bac982b95d8","Type":"ContainerDied","Data":"c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d"} Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.026773 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" event={"ID":"4ade1efd-98f9-4aa8-a415-6bac982b95d8","Type":"ContainerDied","Data":"5fef36905607f1d5580a400444633fc06ffc48286a52391f6464d12cc5809b29"} Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.026776 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-pfcv6" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.026794 4669 scope.go:117] "RemoveContainer" containerID="c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.059514 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-pfcv6"] Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.060747 4669 scope.go:117] "RemoveContainer" containerID="f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.071530 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-pfcv6"] Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.086261 4669 scope.go:117] "RemoveContainer" containerID="c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d" Oct 10 09:32:57 crc kubenswrapper[4669]: E1010 09:32:57.088815 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d\": container with ID starting with c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d not found: ID does not exist" containerID="c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.088862 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d"} err="failed to get container status \"c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d\": rpc error: code = NotFound desc = could not find container \"c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d\": container with ID starting with c3e47521436e797d9f7a4d19f763c5805537483809d44c64326be941cf7d3e6d not found: ID does not exist" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.088890 4669 scope.go:117] "RemoveContainer" containerID="f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97" Oct 10 09:32:57 crc kubenswrapper[4669]: E1010 09:32:57.089322 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97\": container with ID starting with f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97 not found: ID does not exist" containerID="f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.089342 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97"} err="failed to get container status \"f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97\": rpc error: code = NotFound desc = could not find container \"f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97\": container with ID starting with f4bc0d4192475971a5a99a7e8015b7cb002817eb9b99977066f888d598700e97 not found: ID does not exist" Oct 10 09:32:57 crc kubenswrapper[4669]: I1010 09:32:57.810737 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" path="/var/lib/kubelet/pods/4ade1efd-98f9-4aa8-a415-6bac982b95d8/volumes" Oct 10 09:33:05 crc kubenswrapper[4669]: I1010 09:33:05.109984 4669 generic.go:334] "Generic (PLEG): container finished" podID="c1172c8c-d2a5-4f42-b97a-87265385580b" containerID="41748360c429b731cab9fa6e47baa4fdeb44a2b9e5c7c023cfb593301f2a12bf" exitCode=0 Oct 10 09:33:05 crc kubenswrapper[4669]: I1010 09:33:05.110029 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1172c8c-d2a5-4f42-b97a-87265385580b","Type":"ContainerDied","Data":"41748360c429b731cab9fa6e47baa4fdeb44a2b9e5c7c023cfb593301f2a12bf"} Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.051236 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj"] Oct 10 09:33:06 crc kubenswrapper[4669]: E1010 09:33:06.052114 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerName="dnsmasq-dns" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.052135 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerName="dnsmasq-dns" Oct 10 09:33:06 crc kubenswrapper[4669]: E1010 09:33:06.052171 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerName="init" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.052179 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerName="init" Oct 10 09:33:06 crc kubenswrapper[4669]: E1010 09:33:06.052195 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerName="dnsmasq-dns" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.052202 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerName="dnsmasq-dns" Oct 10 09:33:06 crc kubenswrapper[4669]: E1010 09:33:06.052214 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerName="init" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.052222 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerName="init" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.052461 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ade1efd-98f9-4aa8-a415-6bac982b95d8" containerName="dnsmasq-dns" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.052479 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b9f4b3b-8d7c-499f-a32f-8662c0ae873d" containerName="dnsmasq-dns" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.053203 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.056230 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.056807 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.057942 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.062270 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj"] Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.063838 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.121163 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c1172c8c-d2a5-4f42-b97a-87265385580b","Type":"ContainerStarted","Data":"ebfe9d2286ca958a71a6f326f98b325153cd78aa900fccdd01a94c323b822b4b"} Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.122858 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.123863 4669 generic.go:334] "Generic (PLEG): container finished" podID="05416508-674b-44d8-9074-62776915ec92" containerID="ba4e4454caac1c3b0dc597fad0f94916957f952155c50a70eafd78f05dc88d0e" exitCode=0 Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.123891 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05416508-674b-44d8-9074-62776915ec92","Type":"ContainerDied","Data":"ba4e4454caac1c3b0dc597fad0f94916957f952155c50a70eafd78f05dc88d0e"} Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.165281 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.165260430000004 podStartE2EDuration="37.16526043s" podCreationTimestamp="2025-10-10 09:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:33:06.158388469 +0000 UTC m=+1329.174407261" watchObservedRunningTime="2025-10-10 09:33:06.16526043 +0000 UTC m=+1329.181279172" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.237905 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.238068 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25zx9\" (UniqueName: \"kubernetes.io/projected/e253a578-2a65-406b-8b9d-3128c12c2cdb-kube-api-access-25zx9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.238114 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.238227 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.339507 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.339774 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.339911 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25zx9\" (UniqueName: \"kubernetes.io/projected/e253a578-2a65-406b-8b9d-3128c12c2cdb-kube-api-access-25zx9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.339993 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.344559 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.346579 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.346990 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.358943 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25zx9\" (UniqueName: \"kubernetes.io/projected/e253a578-2a65-406b-8b9d-3128c12c2cdb-kube-api-access-25zx9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.374092 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:06 crc kubenswrapper[4669]: I1010 09:33:06.753068 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj"] Oct 10 09:33:06 crc kubenswrapper[4669]: W1010 09:33:06.761518 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode253a578_2a65_406b_8b9d_3128c12c2cdb.slice/crio-0c5c5a629cc3db462c82c26b65d3dc895dd1b56eb85cd6366fe15cf94f8c884a WatchSource:0}: Error finding container 0c5c5a629cc3db462c82c26b65d3dc895dd1b56eb85cd6366fe15cf94f8c884a: Status 404 returned error can't find the container with id 0c5c5a629cc3db462c82c26b65d3dc895dd1b56eb85cd6366fe15cf94f8c884a Oct 10 09:33:07 crc kubenswrapper[4669]: I1010 09:33:07.143245 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05416508-674b-44d8-9074-62776915ec92","Type":"ContainerStarted","Data":"df69dc3c4745a67367c4355c564db6602624e463a762125b7b923e1fd6997965"} Oct 10 09:33:07 crc kubenswrapper[4669]: I1010 09:33:07.143791 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:33:07 crc kubenswrapper[4669]: I1010 09:33:07.144958 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" event={"ID":"e253a578-2a65-406b-8b9d-3128c12c2cdb","Type":"ContainerStarted","Data":"0c5c5a629cc3db462c82c26b65d3dc895dd1b56eb85cd6366fe15cf94f8c884a"} Oct 10 09:33:07 crc kubenswrapper[4669]: I1010 09:33:07.176889 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.176865595 podStartE2EDuration="36.176865595s" podCreationTimestamp="2025-10-10 09:32:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:33:07.167725782 +0000 UTC m=+1330.183744524" watchObservedRunningTime="2025-10-10 09:33:07.176865595 +0000 UTC m=+1330.192884357" Oct 10 09:33:17 crc kubenswrapper[4669]: I1010 09:33:17.238787 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" event={"ID":"e253a578-2a65-406b-8b9d-3128c12c2cdb","Type":"ContainerStarted","Data":"e08600586c8017a0e8c5f7d87bfd6f1c8641110f86a20271453e5878e6a3cb35"} Oct 10 09:33:17 crc kubenswrapper[4669]: I1010 09:33:17.264448 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" podStartSLOduration=1.786552058 podStartE2EDuration="11.264428033s" podCreationTimestamp="2025-10-10 09:33:06 +0000 UTC" firstStartedPulling="2025-10-10 09:33:06.763921158 +0000 UTC m=+1329.779939900" lastFinishedPulling="2025-10-10 09:33:16.241797133 +0000 UTC m=+1339.257815875" observedRunningTime="2025-10-10 09:33:17.261621793 +0000 UTC m=+1340.277640545" watchObservedRunningTime="2025-10-10 09:33:17.264428033 +0000 UTC m=+1340.280446785" Oct 10 09:33:20 crc kubenswrapper[4669]: I1010 09:33:20.340727 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 10 09:33:21 crc kubenswrapper[4669]: I1010 09:33:21.420772 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 10 09:33:24 crc kubenswrapper[4669]: I1010 09:33:24.275236 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:33:24 crc kubenswrapper[4669]: I1010 09:33:24.275650 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:33:24 crc kubenswrapper[4669]: I1010 09:33:24.275711 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:33:24 crc kubenswrapper[4669]: I1010 09:33:24.276482 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dcac214f7000ce8e75b0c4db264661af235fd9f2455e77dacca6bf9bb80e4be0"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:33:24 crc kubenswrapper[4669]: I1010 09:33:24.276547 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://dcac214f7000ce8e75b0c4db264661af235fd9f2455e77dacca6bf9bb80e4be0" gracePeriod=600 Oct 10 09:33:25 crc kubenswrapper[4669]: I1010 09:33:25.328088 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="dcac214f7000ce8e75b0c4db264661af235fd9f2455e77dacca6bf9bb80e4be0" exitCode=0 Oct 10 09:33:25 crc kubenswrapper[4669]: I1010 09:33:25.328285 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"dcac214f7000ce8e75b0c4db264661af235fd9f2455e77dacca6bf9bb80e4be0"} Oct 10 09:33:25 crc kubenswrapper[4669]: I1010 09:33:25.328729 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9"} Oct 10 09:33:25 crc kubenswrapper[4669]: I1010 09:33:25.328756 4669 scope.go:117] "RemoveContainer" containerID="d4c47f9d76180d8158ee4a92da9144583bfce3329aa06b2cf0ab334f9446a4ef" Oct 10 09:33:28 crc kubenswrapper[4669]: I1010 09:33:28.359872 4669 generic.go:334] "Generic (PLEG): container finished" podID="e253a578-2a65-406b-8b9d-3128c12c2cdb" containerID="e08600586c8017a0e8c5f7d87bfd6f1c8641110f86a20271453e5878e6a3cb35" exitCode=0 Oct 10 09:33:28 crc kubenswrapper[4669]: I1010 09:33:28.359998 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" event={"ID":"e253a578-2a65-406b-8b9d-3128c12c2cdb","Type":"ContainerDied","Data":"e08600586c8017a0e8c5f7d87bfd6f1c8641110f86a20271453e5878e6a3cb35"} Oct 10 09:33:29 crc kubenswrapper[4669]: I1010 09:33:29.809456 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.012128 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25zx9\" (UniqueName: \"kubernetes.io/projected/e253a578-2a65-406b-8b9d-3128c12c2cdb-kube-api-access-25zx9\") pod \"e253a578-2a65-406b-8b9d-3128c12c2cdb\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.012375 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-inventory\") pod \"e253a578-2a65-406b-8b9d-3128c12c2cdb\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.012401 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-repo-setup-combined-ca-bundle\") pod \"e253a578-2a65-406b-8b9d-3128c12c2cdb\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.012438 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-ssh-key\") pod \"e253a578-2a65-406b-8b9d-3128c12c2cdb\" (UID: \"e253a578-2a65-406b-8b9d-3128c12c2cdb\") " Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.017859 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e253a578-2a65-406b-8b9d-3128c12c2cdb-kube-api-access-25zx9" (OuterVolumeSpecName: "kube-api-access-25zx9") pod "e253a578-2a65-406b-8b9d-3128c12c2cdb" (UID: "e253a578-2a65-406b-8b9d-3128c12c2cdb"). InnerVolumeSpecName "kube-api-access-25zx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.018905 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e253a578-2a65-406b-8b9d-3128c12c2cdb" (UID: "e253a578-2a65-406b-8b9d-3128c12c2cdb"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.042243 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-inventory" (OuterVolumeSpecName: "inventory") pod "e253a578-2a65-406b-8b9d-3128c12c2cdb" (UID: "e253a578-2a65-406b-8b9d-3128c12c2cdb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.051290 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e253a578-2a65-406b-8b9d-3128c12c2cdb" (UID: "e253a578-2a65-406b-8b9d-3128c12c2cdb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.119784 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25zx9\" (UniqueName: \"kubernetes.io/projected/e253a578-2a65-406b-8b9d-3128c12c2cdb-kube-api-access-25zx9\") on node \"crc\" DevicePath \"\"" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.119828 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.119842 4669 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.119854 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e253a578-2a65-406b-8b9d-3128c12c2cdb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.387066 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" event={"ID":"e253a578-2a65-406b-8b9d-3128c12c2cdb","Type":"ContainerDied","Data":"0c5c5a629cc3db462c82c26b65d3dc895dd1b56eb85cd6366fe15cf94f8c884a"} Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.387451 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c5c5a629cc3db462c82c26b65d3dc895dd1b56eb85cd6366fe15cf94f8c884a" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.387367 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.491062 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp"] Oct 10 09:33:30 crc kubenswrapper[4669]: E1010 09:33:30.491679 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e253a578-2a65-406b-8b9d-3128c12c2cdb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.491805 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e253a578-2a65-406b-8b9d-3128c12c2cdb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.492108 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e253a578-2a65-406b-8b9d-3128c12c2cdb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.492851 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.495369 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.495865 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.496105 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.496823 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.505197 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp"] Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.631670 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.631776 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xpn6\" (UniqueName: \"kubernetes.io/projected/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-kube-api-access-6xpn6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.631822 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.632388 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.733842 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.733906 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.733996 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xpn6\" (UniqueName: \"kubernetes.io/projected/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-kube-api-access-6xpn6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.734025 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.738198 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.746296 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.751056 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.753560 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xpn6\" (UniqueName: \"kubernetes.io/projected/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-kube-api-access-6xpn6\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:30 crc kubenswrapper[4669]: I1010 09:33:30.813037 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:33:31 crc kubenswrapper[4669]: I1010 09:33:31.419448 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp"] Oct 10 09:33:32 crc kubenswrapper[4669]: I1010 09:33:32.404769 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" event={"ID":"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94","Type":"ContainerStarted","Data":"66e1ed52402bd6ce43e45c69541ce9f7a9bdd8fefb2362e54dfc2a36a0ab1f02"} Oct 10 09:33:32 crc kubenswrapper[4669]: I1010 09:33:32.405223 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" event={"ID":"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94","Type":"ContainerStarted","Data":"6636d6b32b6210197d076f6f76a5dac5f48d643dd5d538c5e1f0e1c12322a7f8"} Oct 10 09:33:32 crc kubenswrapper[4669]: I1010 09:33:32.424347 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" podStartSLOduration=1.88379371 podStartE2EDuration="2.424318259s" podCreationTimestamp="2025-10-10 09:33:30 +0000 UTC" firstStartedPulling="2025-10-10 09:33:31.408833649 +0000 UTC m=+1354.424852391" lastFinishedPulling="2025-10-10 09:33:31.949358198 +0000 UTC m=+1354.965376940" observedRunningTime="2025-10-10 09:33:32.419663629 +0000 UTC m=+1355.435682371" watchObservedRunningTime="2025-10-10 09:33:32.424318259 +0000 UTC m=+1355.440337041" Oct 10 09:34:09 crc kubenswrapper[4669]: I1010 09:34:09.301425 4669 scope.go:117] "RemoveContainer" containerID="d37d3fd44e8c91fade589abd164824cb7e60a615a7ec70be22037907b2ee5cb6" Oct 10 09:34:09 crc kubenswrapper[4669]: I1010 09:34:09.376487 4669 scope.go:117] "RemoveContainer" containerID="5a021356d80a083e165c66ee1fb989416744bf9171ad4471c6138eb28958ca9a" Oct 10 09:34:09 crc kubenswrapper[4669]: I1010 09:34:09.430429 4669 scope.go:117] "RemoveContainer" containerID="e5b0a3c74455aca34a9a12e3eafd45bb2cd510b1eae7a325e0ae3ff191bc58e0" Oct 10 09:34:35 crc kubenswrapper[4669]: I1010 09:34:35.988459 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8jrrt"] Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.019560 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.030942 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8jrrt"] Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.147954 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-utilities\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.148263 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxbpr\" (UniqueName: \"kubernetes.io/projected/3feee209-f699-4571-a5c0-548bd40e5d3a-kube-api-access-gxbpr\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.148465 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-catalog-content\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.250124 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-utilities\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.250195 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxbpr\" (UniqueName: \"kubernetes.io/projected/3feee209-f699-4571-a5c0-548bd40e5d3a-kube-api-access-gxbpr\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.250379 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-catalog-content\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.250707 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-utilities\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.250815 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-catalog-content\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.290674 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxbpr\" (UniqueName: \"kubernetes.io/projected/3feee209-f699-4571-a5c0-548bd40e5d3a-kube-api-access-gxbpr\") pod \"community-operators-8jrrt\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:36 crc kubenswrapper[4669]: I1010 09:34:36.361571 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:37 crc kubenswrapper[4669]: I1010 09:34:37.001574 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8jrrt"] Oct 10 09:34:37 crc kubenswrapper[4669]: I1010 09:34:37.006402 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8jrrt" event={"ID":"3feee209-f699-4571-a5c0-548bd40e5d3a","Type":"ContainerStarted","Data":"bf46b16f0d26a86efdcef4d41354bd38f1d20ce1b07423eb8be149a7bef0e4c3"} Oct 10 09:34:38 crc kubenswrapper[4669]: I1010 09:34:38.015937 4669 generic.go:334] "Generic (PLEG): container finished" podID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerID="4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7" exitCode=0 Oct 10 09:34:38 crc kubenswrapper[4669]: I1010 09:34:38.016020 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8jrrt" event={"ID":"3feee209-f699-4571-a5c0-548bd40e5d3a","Type":"ContainerDied","Data":"4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7"} Oct 10 09:34:40 crc kubenswrapper[4669]: I1010 09:34:40.035915 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8jrrt" event={"ID":"3feee209-f699-4571-a5c0-548bd40e5d3a","Type":"ContainerStarted","Data":"3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e"} Oct 10 09:34:41 crc kubenswrapper[4669]: I1010 09:34:41.050147 4669 generic.go:334] "Generic (PLEG): container finished" podID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerID="3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e" exitCode=0 Oct 10 09:34:41 crc kubenswrapper[4669]: I1010 09:34:41.050267 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8jrrt" event={"ID":"3feee209-f699-4571-a5c0-548bd40e5d3a","Type":"ContainerDied","Data":"3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e"} Oct 10 09:34:42 crc kubenswrapper[4669]: I1010 09:34:42.060820 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8jrrt" event={"ID":"3feee209-f699-4571-a5c0-548bd40e5d3a","Type":"ContainerStarted","Data":"cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5"} Oct 10 09:34:42 crc kubenswrapper[4669]: I1010 09:34:42.079644 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8jrrt" podStartSLOduration=3.423043223 podStartE2EDuration="7.079625154s" podCreationTimestamp="2025-10-10 09:34:35 +0000 UTC" firstStartedPulling="2025-10-10 09:34:38.018317692 +0000 UTC m=+1421.034336434" lastFinishedPulling="2025-10-10 09:34:41.674899623 +0000 UTC m=+1424.690918365" observedRunningTime="2025-10-10 09:34:42.077166005 +0000 UTC m=+1425.093184737" watchObservedRunningTime="2025-10-10 09:34:42.079625154 +0000 UTC m=+1425.095643896" Oct 10 09:34:46 crc kubenswrapper[4669]: I1010 09:34:46.362361 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:46 crc kubenswrapper[4669]: I1010 09:34:46.362907 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:46 crc kubenswrapper[4669]: I1010 09:34:46.417971 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:47 crc kubenswrapper[4669]: I1010 09:34:47.172057 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:47 crc kubenswrapper[4669]: I1010 09:34:47.231089 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8jrrt"] Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.136470 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8jrrt" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="registry-server" containerID="cri-o://cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5" gracePeriod=2 Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.681067 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.808503 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxbpr\" (UniqueName: \"kubernetes.io/projected/3feee209-f699-4571-a5c0-548bd40e5d3a-kube-api-access-gxbpr\") pod \"3feee209-f699-4571-a5c0-548bd40e5d3a\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.808561 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-utilities\") pod \"3feee209-f699-4571-a5c0-548bd40e5d3a\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.808669 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-catalog-content\") pod \"3feee209-f699-4571-a5c0-548bd40e5d3a\" (UID: \"3feee209-f699-4571-a5c0-548bd40e5d3a\") " Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.811985 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-utilities" (OuterVolumeSpecName: "utilities") pod "3feee209-f699-4571-a5c0-548bd40e5d3a" (UID: "3feee209-f699-4571-a5c0-548bd40e5d3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.826014 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3feee209-f699-4571-a5c0-548bd40e5d3a-kube-api-access-gxbpr" (OuterVolumeSpecName: "kube-api-access-gxbpr") pod "3feee209-f699-4571-a5c0-548bd40e5d3a" (UID: "3feee209-f699-4571-a5c0-548bd40e5d3a"). InnerVolumeSpecName "kube-api-access-gxbpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.876243 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3feee209-f699-4571-a5c0-548bd40e5d3a" (UID: "3feee209-f699-4571-a5c0-548bd40e5d3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.911217 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxbpr\" (UniqueName: \"kubernetes.io/projected/3feee209-f699-4571-a5c0-548bd40e5d3a-kube-api-access-gxbpr\") on node \"crc\" DevicePath \"\"" Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.911252 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:34:49 crc kubenswrapper[4669]: I1010 09:34:49.911267 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feee209-f699-4571-a5c0-548bd40e5d3a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.151489 4669 generic.go:334] "Generic (PLEG): container finished" podID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerID="cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5" exitCode=0 Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.151539 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8jrrt" event={"ID":"3feee209-f699-4571-a5c0-548bd40e5d3a","Type":"ContainerDied","Data":"cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5"} Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.151574 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8jrrt" event={"ID":"3feee209-f699-4571-a5c0-548bd40e5d3a","Type":"ContainerDied","Data":"bf46b16f0d26a86efdcef4d41354bd38f1d20ce1b07423eb8be149a7bef0e4c3"} Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.151620 4669 scope.go:117] "RemoveContainer" containerID="cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.151621 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8jrrt" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.184802 4669 scope.go:117] "RemoveContainer" containerID="3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.220872 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8jrrt"] Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.229227 4669 scope.go:117] "RemoveContainer" containerID="4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.231936 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8jrrt"] Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.286391 4669 scope.go:117] "RemoveContainer" containerID="cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5" Oct 10 09:34:50 crc kubenswrapper[4669]: E1010 09:34:50.286918 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5\": container with ID starting with cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5 not found: ID does not exist" containerID="cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.286960 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5"} err="failed to get container status \"cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5\": rpc error: code = NotFound desc = could not find container \"cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5\": container with ID starting with cbd7b1f6e6df272c97ac7274c9175be5ac23fcb2ba51f6e151b94c766605c4a5 not found: ID does not exist" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.286988 4669 scope.go:117] "RemoveContainer" containerID="3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e" Oct 10 09:34:50 crc kubenswrapper[4669]: E1010 09:34:50.287213 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e\": container with ID starting with 3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e not found: ID does not exist" containerID="3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.287243 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e"} err="failed to get container status \"3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e\": rpc error: code = NotFound desc = could not find container \"3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e\": container with ID starting with 3321da203bd307ffafc9a8dd30c5f8d155439dfe5e2f46899a41258a3be2c00e not found: ID does not exist" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.287264 4669 scope.go:117] "RemoveContainer" containerID="4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7" Oct 10 09:34:50 crc kubenswrapper[4669]: E1010 09:34:50.287475 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7\": container with ID starting with 4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7 not found: ID does not exist" containerID="4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7" Oct 10 09:34:50 crc kubenswrapper[4669]: I1010 09:34:50.287505 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7"} err="failed to get container status \"4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7\": rpc error: code = NotFound desc = could not find container \"4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7\": container with ID starting with 4aa55eb7a7a67af138214db36649f9af6868c28a9df0552029571c5e1fba6ee7 not found: ID does not exist" Oct 10 09:34:51 crc kubenswrapper[4669]: I1010 09:34:51.815675 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" path="/var/lib/kubelet/pods/3feee209-f699-4571-a5c0-548bd40e5d3a/volumes" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.456508 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p9gm7"] Oct 10 09:35:05 crc kubenswrapper[4669]: E1010 09:35:05.457474 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="registry-server" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.457492 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="registry-server" Oct 10 09:35:05 crc kubenswrapper[4669]: E1010 09:35:05.457517 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="extract-utilities" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.457525 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="extract-utilities" Oct 10 09:35:05 crc kubenswrapper[4669]: E1010 09:35:05.457559 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="extract-content" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.457568 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="extract-content" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.457850 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feee209-f699-4571-a5c0-548bd40e5d3a" containerName="registry-server" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.459517 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.471913 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9gm7"] Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.500057 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-catalog-content\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.500110 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfl2w\" (UniqueName: \"kubernetes.io/projected/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-kube-api-access-bfl2w\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.500159 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-utilities\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.601716 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-catalog-content\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.601768 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfl2w\" (UniqueName: \"kubernetes.io/projected/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-kube-api-access-bfl2w\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.601835 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-utilities\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.602812 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-utilities\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.603077 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-catalog-content\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.629928 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfl2w\" (UniqueName: \"kubernetes.io/projected/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-kube-api-access-bfl2w\") pod \"redhat-marketplace-p9gm7\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:05 crc kubenswrapper[4669]: I1010 09:35:05.780544 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:06 crc kubenswrapper[4669]: I1010 09:35:06.250565 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9gm7"] Oct 10 09:35:06 crc kubenswrapper[4669]: I1010 09:35:06.308433 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9gm7" event={"ID":"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba","Type":"ContainerStarted","Data":"26042e2c2c481cd06d937a0cd68f9ab3bdee8dfec18bee7f4ef68a332888eb6c"} Oct 10 09:35:07 crc kubenswrapper[4669]: I1010 09:35:07.338438 4669 generic.go:334] "Generic (PLEG): container finished" podID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerID="d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3" exitCode=0 Oct 10 09:35:07 crc kubenswrapper[4669]: I1010 09:35:07.338485 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9gm7" event={"ID":"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba","Type":"ContainerDied","Data":"d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3"} Oct 10 09:35:08 crc kubenswrapper[4669]: I1010 09:35:08.350513 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9gm7" event={"ID":"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba","Type":"ContainerStarted","Data":"13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d"} Oct 10 09:35:09 crc kubenswrapper[4669]: I1010 09:35:09.364562 4669 generic.go:334] "Generic (PLEG): container finished" podID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerID="13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d" exitCode=0 Oct 10 09:35:09 crc kubenswrapper[4669]: I1010 09:35:09.364692 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9gm7" event={"ID":"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba","Type":"ContainerDied","Data":"13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d"} Oct 10 09:35:10 crc kubenswrapper[4669]: I1010 09:35:10.375075 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9gm7" event={"ID":"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba","Type":"ContainerStarted","Data":"bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50"} Oct 10 09:35:10 crc kubenswrapper[4669]: I1010 09:35:10.394783 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p9gm7" podStartSLOduration=2.789798107 podStartE2EDuration="5.394765442s" podCreationTimestamp="2025-10-10 09:35:05 +0000 UTC" firstStartedPulling="2025-10-10 09:35:07.342031537 +0000 UTC m=+1450.358050279" lastFinishedPulling="2025-10-10 09:35:09.946998872 +0000 UTC m=+1452.963017614" observedRunningTime="2025-10-10 09:35:10.392120178 +0000 UTC m=+1453.408138910" watchObservedRunningTime="2025-10-10 09:35:10.394765442 +0000 UTC m=+1453.410784184" Oct 10 09:35:15 crc kubenswrapper[4669]: I1010 09:35:15.782328 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:15 crc kubenswrapper[4669]: I1010 09:35:15.782977 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:15 crc kubenswrapper[4669]: I1010 09:35:15.901922 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:16 crc kubenswrapper[4669]: I1010 09:35:16.479571 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:16 crc kubenswrapper[4669]: I1010 09:35:16.526078 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9gm7"] Oct 10 09:35:18 crc kubenswrapper[4669]: I1010 09:35:18.454155 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p9gm7" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="registry-server" containerID="cri-o://bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50" gracePeriod=2 Oct 10 09:35:18 crc kubenswrapper[4669]: I1010 09:35:18.892072 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:18 crc kubenswrapper[4669]: I1010 09:35:18.999758 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-utilities\") pod \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " Oct 10 09:35:18 crc kubenswrapper[4669]: I1010 09:35:18.999845 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-catalog-content\") pod \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " Oct 10 09:35:18 crc kubenswrapper[4669]: I1010 09:35:18.999913 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfl2w\" (UniqueName: \"kubernetes.io/projected/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-kube-api-access-bfl2w\") pod \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\" (UID: \"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba\") " Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.001498 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-utilities" (OuterVolumeSpecName: "utilities") pod "9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" (UID: "9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.009006 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-kube-api-access-bfl2w" (OuterVolumeSpecName: "kube-api-access-bfl2w") pod "9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" (UID: "9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba"). InnerVolumeSpecName "kube-api-access-bfl2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.014533 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" (UID: "9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.101373 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfl2w\" (UniqueName: \"kubernetes.io/projected/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-kube-api-access-bfl2w\") on node \"crc\" DevicePath \"\"" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.101410 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.101420 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.471886 4669 generic.go:334] "Generic (PLEG): container finished" podID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerID="bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50" exitCode=0 Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.471922 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9gm7" event={"ID":"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba","Type":"ContainerDied","Data":"bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50"} Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.471962 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p9gm7" event={"ID":"9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba","Type":"ContainerDied","Data":"26042e2c2c481cd06d937a0cd68f9ab3bdee8dfec18bee7f4ef68a332888eb6c"} Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.471979 4669 scope.go:117] "RemoveContainer" containerID="bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.471985 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p9gm7" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.505166 4669 scope.go:117] "RemoveContainer" containerID="13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.526799 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9gm7"] Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.528973 4669 scope.go:117] "RemoveContainer" containerID="d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.536676 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p9gm7"] Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.568889 4669 scope.go:117] "RemoveContainer" containerID="bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50" Oct 10 09:35:19 crc kubenswrapper[4669]: E1010 09:35:19.569308 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50\": container with ID starting with bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50 not found: ID does not exist" containerID="bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.569338 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50"} err="failed to get container status \"bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50\": rpc error: code = NotFound desc = could not find container \"bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50\": container with ID starting with bf4c08a3c48b8dc69e4ddfd6483b37c44e19929ac7dabbfc3b35bb86c2142a50 not found: ID does not exist" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.569462 4669 scope.go:117] "RemoveContainer" containerID="13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d" Oct 10 09:35:19 crc kubenswrapper[4669]: E1010 09:35:19.569824 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d\": container with ID starting with 13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d not found: ID does not exist" containerID="13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.569853 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d"} err="failed to get container status \"13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d\": rpc error: code = NotFound desc = could not find container \"13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d\": container with ID starting with 13bf7f0955ddafcf9cce8dbcdf66577ca1ae92faefa0199f1d7fb8384826a92d not found: ID does not exist" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.569870 4669 scope.go:117] "RemoveContainer" containerID="d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3" Oct 10 09:35:19 crc kubenswrapper[4669]: E1010 09:35:19.570553 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3\": container with ID starting with d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3 not found: ID does not exist" containerID="d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.570594 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3"} err="failed to get container status \"d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3\": rpc error: code = NotFound desc = could not find container \"d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3\": container with ID starting with d4a754fc280a239a961a8348d98d7d11952f2f659efa53becdee2cefb89c8ed3 not found: ID does not exist" Oct 10 09:35:19 crc kubenswrapper[4669]: I1010 09:35:19.810732 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" path="/var/lib/kubelet/pods/9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba/volumes" Oct 10 09:35:24 crc kubenswrapper[4669]: I1010 09:35:24.275051 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:35:24 crc kubenswrapper[4669]: I1010 09:35:24.275137 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.376745 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w29sv"] Oct 10 09:35:38 crc kubenswrapper[4669]: E1010 09:35:38.379303 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="extract-utilities" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.379321 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="extract-utilities" Oct 10 09:35:38 crc kubenswrapper[4669]: E1010 09:35:38.379343 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="extract-content" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.379349 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="extract-content" Oct 10 09:35:38 crc kubenswrapper[4669]: E1010 09:35:38.379367 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="registry-server" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.379375 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="registry-server" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.390430 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aaa12a0-f4e6-4a6c-9c55-172d9028e8ba" containerName="registry-server" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.392225 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w29sv"] Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.392309 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.484808 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qpd4\" (UniqueName: \"kubernetes.io/projected/8c6d51a0-df97-4864-af87-75d6463df34a-kube-api-access-9qpd4\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.484853 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-catalog-content\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.484915 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-utilities\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.586042 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qpd4\" (UniqueName: \"kubernetes.io/projected/8c6d51a0-df97-4864-af87-75d6463df34a-kube-api-access-9qpd4\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.586100 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-catalog-content\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.586162 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-utilities\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.586764 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-utilities\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.587330 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-catalog-content\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.614518 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qpd4\" (UniqueName: \"kubernetes.io/projected/8c6d51a0-df97-4864-af87-75d6463df34a-kube-api-access-9qpd4\") pod \"redhat-operators-w29sv\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:38 crc kubenswrapper[4669]: I1010 09:35:38.721984 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:39 crc kubenswrapper[4669]: I1010 09:35:39.265531 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w29sv"] Oct 10 09:35:39 crc kubenswrapper[4669]: I1010 09:35:39.696047 4669 generic.go:334] "Generic (PLEG): container finished" podID="8c6d51a0-df97-4864-af87-75d6463df34a" containerID="7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874" exitCode=0 Oct 10 09:35:39 crc kubenswrapper[4669]: I1010 09:35:39.696127 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w29sv" event={"ID":"8c6d51a0-df97-4864-af87-75d6463df34a","Type":"ContainerDied","Data":"7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874"} Oct 10 09:35:39 crc kubenswrapper[4669]: I1010 09:35:39.696207 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w29sv" event={"ID":"8c6d51a0-df97-4864-af87-75d6463df34a","Type":"ContainerStarted","Data":"7387b6875a0c1df05d06f5df239d44d006e38de5dda3e498a281aabec00aeabb"} Oct 10 09:35:40 crc kubenswrapper[4669]: I1010 09:35:40.704879 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w29sv" event={"ID":"8c6d51a0-df97-4864-af87-75d6463df34a","Type":"ContainerStarted","Data":"d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd"} Oct 10 09:35:44 crc kubenswrapper[4669]: I1010 09:35:44.744084 4669 generic.go:334] "Generic (PLEG): container finished" podID="8c6d51a0-df97-4864-af87-75d6463df34a" containerID="d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd" exitCode=0 Oct 10 09:35:44 crc kubenswrapper[4669]: I1010 09:35:44.744164 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w29sv" event={"ID":"8c6d51a0-df97-4864-af87-75d6463df34a","Type":"ContainerDied","Data":"d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd"} Oct 10 09:35:45 crc kubenswrapper[4669]: I1010 09:35:45.754836 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w29sv" event={"ID":"8c6d51a0-df97-4864-af87-75d6463df34a","Type":"ContainerStarted","Data":"191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1"} Oct 10 09:35:45 crc kubenswrapper[4669]: I1010 09:35:45.776560 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w29sv" podStartSLOduration=2.171717843 podStartE2EDuration="7.776541483s" podCreationTimestamp="2025-10-10 09:35:38 +0000 UTC" firstStartedPulling="2025-10-10 09:35:39.699312193 +0000 UTC m=+1482.715330935" lastFinishedPulling="2025-10-10 09:35:45.304135833 +0000 UTC m=+1488.320154575" observedRunningTime="2025-10-10 09:35:45.773203385 +0000 UTC m=+1488.789222147" watchObservedRunningTime="2025-10-10 09:35:45.776541483 +0000 UTC m=+1488.792560225" Oct 10 09:35:48 crc kubenswrapper[4669]: I1010 09:35:48.722625 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:48 crc kubenswrapper[4669]: I1010 09:35:48.723966 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:49 crc kubenswrapper[4669]: I1010 09:35:49.768474 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w29sv" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="registry-server" probeResult="failure" output=< Oct 10 09:35:49 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 09:35:49 crc kubenswrapper[4669]: > Oct 10 09:35:54 crc kubenswrapper[4669]: I1010 09:35:54.275229 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:35:54 crc kubenswrapper[4669]: I1010 09:35:54.275612 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:35:58 crc kubenswrapper[4669]: I1010 09:35:58.783838 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:58 crc kubenswrapper[4669]: I1010 09:35:58.843670 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:35:59 crc kubenswrapper[4669]: I1010 09:35:59.028361 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w29sv"] Oct 10 09:35:59 crc kubenswrapper[4669]: I1010 09:35:59.893541 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w29sv" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="registry-server" containerID="cri-o://191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1" gracePeriod=2 Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.345496 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.393489 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-utilities\") pod \"8c6d51a0-df97-4864-af87-75d6463df34a\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.393565 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-catalog-content\") pod \"8c6d51a0-df97-4864-af87-75d6463df34a\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.397461 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-utilities" (OuterVolumeSpecName: "utilities") pod "8c6d51a0-df97-4864-af87-75d6463df34a" (UID: "8c6d51a0-df97-4864-af87-75d6463df34a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.487964 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c6d51a0-df97-4864-af87-75d6463df34a" (UID: "8c6d51a0-df97-4864-af87-75d6463df34a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.495464 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qpd4\" (UniqueName: \"kubernetes.io/projected/8c6d51a0-df97-4864-af87-75d6463df34a-kube-api-access-9qpd4\") pod \"8c6d51a0-df97-4864-af87-75d6463df34a\" (UID: \"8c6d51a0-df97-4864-af87-75d6463df34a\") " Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.496170 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.496195 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c6d51a0-df97-4864-af87-75d6463df34a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.500701 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c6d51a0-df97-4864-af87-75d6463df34a-kube-api-access-9qpd4" (OuterVolumeSpecName: "kube-api-access-9qpd4") pod "8c6d51a0-df97-4864-af87-75d6463df34a" (UID: "8c6d51a0-df97-4864-af87-75d6463df34a"). InnerVolumeSpecName "kube-api-access-9qpd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.597515 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qpd4\" (UniqueName: \"kubernetes.io/projected/8c6d51a0-df97-4864-af87-75d6463df34a-kube-api-access-9qpd4\") on node \"crc\" DevicePath \"\"" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.905866 4669 generic.go:334] "Generic (PLEG): container finished" podID="8c6d51a0-df97-4864-af87-75d6463df34a" containerID="191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1" exitCode=0 Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.905909 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w29sv" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.905916 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w29sv" event={"ID":"8c6d51a0-df97-4864-af87-75d6463df34a","Type":"ContainerDied","Data":"191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1"} Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.905942 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w29sv" event={"ID":"8c6d51a0-df97-4864-af87-75d6463df34a","Type":"ContainerDied","Data":"7387b6875a0c1df05d06f5df239d44d006e38de5dda3e498a281aabec00aeabb"} Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.905959 4669 scope.go:117] "RemoveContainer" containerID="191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.937730 4669 scope.go:117] "RemoveContainer" containerID="d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd" Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.943126 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w29sv"] Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.961483 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w29sv"] Oct 10 09:36:00 crc kubenswrapper[4669]: I1010 09:36:00.980417 4669 scope.go:117] "RemoveContainer" containerID="7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874" Oct 10 09:36:01 crc kubenswrapper[4669]: I1010 09:36:01.021465 4669 scope.go:117] "RemoveContainer" containerID="191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1" Oct 10 09:36:01 crc kubenswrapper[4669]: E1010 09:36:01.021997 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1\": container with ID starting with 191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1 not found: ID does not exist" containerID="191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1" Oct 10 09:36:01 crc kubenswrapper[4669]: I1010 09:36:01.022037 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1"} err="failed to get container status \"191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1\": rpc error: code = NotFound desc = could not find container \"191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1\": container with ID starting with 191e74783c2dde0cc1d1e3394e556f9c89c5074369dc174d23f83fd8370020c1 not found: ID does not exist" Oct 10 09:36:01 crc kubenswrapper[4669]: I1010 09:36:01.022067 4669 scope.go:117] "RemoveContainer" containerID="d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd" Oct 10 09:36:01 crc kubenswrapper[4669]: E1010 09:36:01.022393 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd\": container with ID starting with d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd not found: ID does not exist" containerID="d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd" Oct 10 09:36:01 crc kubenswrapper[4669]: I1010 09:36:01.022423 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd"} err="failed to get container status \"d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd\": rpc error: code = NotFound desc = could not find container \"d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd\": container with ID starting with d3e08373f30aacf20b6bf998033be14c1f42821aad924fedb61ec092f97df1cd not found: ID does not exist" Oct 10 09:36:01 crc kubenswrapper[4669]: I1010 09:36:01.022442 4669 scope.go:117] "RemoveContainer" containerID="7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874" Oct 10 09:36:01 crc kubenswrapper[4669]: E1010 09:36:01.023401 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874\": container with ID starting with 7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874 not found: ID does not exist" containerID="7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874" Oct 10 09:36:01 crc kubenswrapper[4669]: I1010 09:36:01.023431 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874"} err="failed to get container status \"7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874\": rpc error: code = NotFound desc = could not find container \"7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874\": container with ID starting with 7dc0b6dd37c4c3d744a5dc4531ae366395a53577b1cc4324572615a66435f874 not found: ID does not exist" Oct 10 09:36:01 crc kubenswrapper[4669]: I1010 09:36:01.808533 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" path="/var/lib/kubelet/pods/8c6d51a0-df97-4864-af87-75d6463df34a/volumes" Oct 10 09:36:09 crc kubenswrapper[4669]: I1010 09:36:09.583807 4669 scope.go:117] "RemoveContainer" containerID="5d47a5af6d1976c9d9facb18d02ab2e18fd1caaf27d5ac75b655fa81b0e2c026" Oct 10 09:36:09 crc kubenswrapper[4669]: I1010 09:36:09.603251 4669 scope.go:117] "RemoveContainer" containerID="3602cf4c842d805bd0a9417942f90453ed004b7f64d3546f7e0d39e0f03c0c0e" Oct 10 09:36:09 crc kubenswrapper[4669]: I1010 09:36:09.627037 4669 scope.go:117] "RemoveContainer" containerID="6ffbb512db5dd9811cd3e49dc36f604e2b55664f85852dc2566433a0b811e804" Oct 10 09:36:09 crc kubenswrapper[4669]: I1010 09:36:09.652402 4669 scope.go:117] "RemoveContainer" containerID="94344d0012806131e11f4312e1b530b37a4c1598cf1f72d7bde7bcd754da4676" Oct 10 09:36:24 crc kubenswrapper[4669]: I1010 09:36:24.275309 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:36:24 crc kubenswrapper[4669]: I1010 09:36:24.275833 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:36:24 crc kubenswrapper[4669]: I1010 09:36:24.275885 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:36:24 crc kubenswrapper[4669]: I1010 09:36:24.276502 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:36:24 crc kubenswrapper[4669]: I1010 09:36:24.276576 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" gracePeriod=600 Oct 10 09:36:24 crc kubenswrapper[4669]: E1010 09:36:24.419991 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:36:25 crc kubenswrapper[4669]: I1010 09:36:25.139569 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" exitCode=0 Oct 10 09:36:25 crc kubenswrapper[4669]: I1010 09:36:25.139627 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9"} Oct 10 09:36:25 crc kubenswrapper[4669]: I1010 09:36:25.139701 4669 scope.go:117] "RemoveContainer" containerID="dcac214f7000ce8e75b0c4db264661af235fd9f2455e77dacca6bf9bb80e4be0" Oct 10 09:36:25 crc kubenswrapper[4669]: I1010 09:36:25.141272 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:36:25 crc kubenswrapper[4669]: E1010 09:36:25.142257 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:36:37 crc kubenswrapper[4669]: I1010 09:36:37.802726 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:36:37 crc kubenswrapper[4669]: E1010 09:36:37.803684 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:36:48 crc kubenswrapper[4669]: I1010 09:36:48.797093 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:36:48 crc kubenswrapper[4669]: E1010 09:36:48.797986 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:36:57 crc kubenswrapper[4669]: I1010 09:36:57.433927 4669 generic.go:334] "Generic (PLEG): container finished" podID="87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" containerID="66e1ed52402bd6ce43e45c69541ce9f7a9bdd8fefb2362e54dfc2a36a0ab1f02" exitCode=0 Oct 10 09:36:57 crc kubenswrapper[4669]: I1010 09:36:57.433994 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" event={"ID":"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94","Type":"ContainerDied","Data":"66e1ed52402bd6ce43e45c69541ce9f7a9bdd8fefb2362e54dfc2a36a0ab1f02"} Oct 10 09:36:58 crc kubenswrapper[4669]: I1010 09:36:58.841465 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.021018 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-ssh-key\") pod \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.021305 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-bootstrap-combined-ca-bundle\") pod \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.021513 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-inventory\") pod \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.021662 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xpn6\" (UniqueName: \"kubernetes.io/projected/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-kube-api-access-6xpn6\") pod \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\" (UID: \"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94\") " Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.026315 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-kube-api-access-6xpn6" (OuterVolumeSpecName: "kube-api-access-6xpn6") pod "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" (UID: "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94"). InnerVolumeSpecName "kube-api-access-6xpn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.029487 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" (UID: "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.048452 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" (UID: "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.052918 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-inventory" (OuterVolumeSpecName: "inventory") pod "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" (UID: "87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.127893 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xpn6\" (UniqueName: \"kubernetes.io/projected/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-kube-api-access-6xpn6\") on node \"crc\" DevicePath \"\"" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.128009 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.128062 4669 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.128111 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.457842 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" event={"ID":"87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94","Type":"ContainerDied","Data":"6636d6b32b6210197d076f6f76a5dac5f48d643dd5d538c5e1f0e1c12322a7f8"} Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.457905 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6636d6b32b6210197d076f6f76a5dac5f48d643dd5d538c5e1f0e1c12322a7f8" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.457938 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.555305 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l"] Oct 10 09:36:59 crc kubenswrapper[4669]: E1010 09:36:59.555774 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="extract-utilities" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.555797 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="extract-utilities" Oct 10 09:36:59 crc kubenswrapper[4669]: E1010 09:36:59.555810 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.555819 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 09:36:59 crc kubenswrapper[4669]: E1010 09:36:59.555833 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="extract-content" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.555841 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="extract-content" Oct 10 09:36:59 crc kubenswrapper[4669]: E1010 09:36:59.555870 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="registry-server" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.555879 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="registry-server" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.556076 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c6d51a0-df97-4864-af87-75d6463df34a" containerName="registry-server" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.556122 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.556796 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.562510 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.562542 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.562562 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.562493 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.577560 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l"] Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.738209 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5ksg\" (UniqueName: \"kubernetes.io/projected/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-kube-api-access-l5ksg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.738400 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.738769 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.795989 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:36:59 crc kubenswrapper[4669]: E1010 09:36:59.796627 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.840861 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.840930 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5ksg\" (UniqueName: \"kubernetes.io/projected/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-kube-api-access-l5ksg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.840977 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.845051 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.845042 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.869486 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5ksg\" (UniqueName: \"kubernetes.io/projected/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-kube-api-access-l5ksg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:36:59 crc kubenswrapper[4669]: I1010 09:36:59.875534 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:37:00 crc kubenswrapper[4669]: I1010 09:37:00.428991 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l"] Oct 10 09:37:00 crc kubenswrapper[4669]: I1010 09:37:00.441742 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:37:00 crc kubenswrapper[4669]: I1010 09:37:00.473104 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" event={"ID":"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148","Type":"ContainerStarted","Data":"f0145a1a774b532e74d4b8fabc4461238b2d75c1cd40a2bd1a1fc28e775da258"} Oct 10 09:37:01 crc kubenswrapper[4669]: I1010 09:37:01.484799 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" event={"ID":"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148","Type":"ContainerStarted","Data":"a9e6be0de9ee8269114d300be5825c5e168fd33df286bfb9dbe3d6b3dd594302"} Oct 10 09:37:01 crc kubenswrapper[4669]: I1010 09:37:01.507482 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" podStartSLOduration=1.882052568 podStartE2EDuration="2.507455992s" podCreationTimestamp="2025-10-10 09:36:59 +0000 UTC" firstStartedPulling="2025-10-10 09:37:00.441433917 +0000 UTC m=+1563.457452659" lastFinishedPulling="2025-10-10 09:37:01.066837331 +0000 UTC m=+1564.082856083" observedRunningTime="2025-10-10 09:37:01.505115557 +0000 UTC m=+1564.521134339" watchObservedRunningTime="2025-10-10 09:37:01.507455992 +0000 UTC m=+1564.523474774" Oct 10 09:37:09 crc kubenswrapper[4669]: I1010 09:37:09.732426 4669 scope.go:117] "RemoveContainer" containerID="202bf293457faa36e046bce932af6e750f2bbb5c297dfe57f8d63ffc21b6e0be" Oct 10 09:37:09 crc kubenswrapper[4669]: I1010 09:37:09.763064 4669 scope.go:117] "RemoveContainer" containerID="cf98206075bede22a040c201cc3d0a2111f3715fc14c2ba83e06b94d116d4a43" Oct 10 09:37:12 crc kubenswrapper[4669]: I1010 09:37:12.795702 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:37:12 crc kubenswrapper[4669]: E1010 09:37:12.796416 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:37:27 crc kubenswrapper[4669]: I1010 09:37:27.805056 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:37:27 crc kubenswrapper[4669]: E1010 09:37:27.806092 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:37:29 crc kubenswrapper[4669]: I1010 09:37:29.048948 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-l8tnz"] Oct 10 09:37:29 crc kubenswrapper[4669]: I1010 09:37:29.057961 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-x8kx4"] Oct 10 09:37:29 crc kubenswrapper[4669]: I1010 09:37:29.065383 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-x8kx4"] Oct 10 09:37:29 crc kubenswrapper[4669]: I1010 09:37:29.073281 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-l8tnz"] Oct 10 09:37:29 crc kubenswrapper[4669]: I1010 09:37:29.809903 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ae32e28-6814-4693-b3e0-bdf5d97f9744" path="/var/lib/kubelet/pods/2ae32e28-6814-4693-b3e0-bdf5d97f9744/volumes" Oct 10 09:37:29 crc kubenswrapper[4669]: I1010 09:37:29.810547 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c" path="/var/lib/kubelet/pods/30aa5ac3-da2c-4bc6-a82d-cfaab0d0fb8c/volumes" Oct 10 09:37:30 crc kubenswrapper[4669]: I1010 09:37:30.037794 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8gt4t"] Oct 10 09:37:30 crc kubenswrapper[4669]: I1010 09:37:30.045542 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8gt4t"] Oct 10 09:37:31 crc kubenswrapper[4669]: I1010 09:37:31.837058 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913a206c-349d-4125-a118-1798385a3d1a" path="/var/lib/kubelet/pods/913a206c-349d-4125-a118-1798385a3d1a/volumes" Oct 10 09:37:39 crc kubenswrapper[4669]: I1010 09:37:39.041072 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2824-account-create-j2pd4"] Oct 10 09:37:39 crc kubenswrapper[4669]: I1010 09:37:39.054816 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2824-account-create-j2pd4"] Oct 10 09:37:39 crc kubenswrapper[4669]: I1010 09:37:39.809735 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d33626d7-d6dd-45c0-8442-1617295ec697" path="/var/lib/kubelet/pods/d33626d7-d6dd-45c0-8442-1617295ec697/volumes" Oct 10 09:37:40 crc kubenswrapper[4669]: I1010 09:37:40.030955 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-30cc-account-create-5h92x"] Oct 10 09:37:40 crc kubenswrapper[4669]: I1010 09:37:40.041743 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ca43-account-create-ts8dc"] Oct 10 09:37:40 crc kubenswrapper[4669]: I1010 09:37:40.049017 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-30cc-account-create-5h92x"] Oct 10 09:37:40 crc kubenswrapper[4669]: I1010 09:37:40.057383 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ca43-account-create-ts8dc"] Oct 10 09:37:40 crc kubenswrapper[4669]: I1010 09:37:40.795506 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:37:40 crc kubenswrapper[4669]: E1010 09:37:40.796725 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:37:41 crc kubenswrapper[4669]: I1010 09:37:41.806543 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a1f32d-7cb9-403c-91aa-28f749dbbc9a" path="/var/lib/kubelet/pods/00a1f32d-7cb9-403c-91aa-28f749dbbc9a/volumes" Oct 10 09:37:41 crc kubenswrapper[4669]: I1010 09:37:41.807190 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeca1748-8fbf-4502-acb8-f1a2821a5c2b" path="/var/lib/kubelet/pods/aeca1748-8fbf-4502-acb8-f1a2821a5c2b/volumes" Oct 10 09:37:53 crc kubenswrapper[4669]: I1010 09:37:53.796349 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:37:53 crc kubenswrapper[4669]: E1010 09:37:53.797579 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:38:02 crc kubenswrapper[4669]: I1010 09:38:02.040085 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8nbn8"] Oct 10 09:38:02 crc kubenswrapper[4669]: I1010 09:38:02.051964 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-tkg49"] Oct 10 09:38:02 crc kubenswrapper[4669]: I1010 09:38:02.068258 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-hxt77"] Oct 10 09:38:02 crc kubenswrapper[4669]: I1010 09:38:02.074877 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-tkg49"] Oct 10 09:38:02 crc kubenswrapper[4669]: I1010 09:38:02.081442 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-hxt77"] Oct 10 09:38:02 crc kubenswrapper[4669]: I1010 09:38:02.088859 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8nbn8"] Oct 10 09:38:03 crc kubenswrapper[4669]: I1010 09:38:03.807011 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c88504-7493-485a-a2ed-e0e62fb4543f" path="/var/lib/kubelet/pods/93c88504-7493-485a-a2ed-e0e62fb4543f/volumes" Oct 10 09:38:03 crc kubenswrapper[4669]: I1010 09:38:03.808114 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="964cd051-6afe-4834-a7b8-ccbe2c547f4e" path="/var/lib/kubelet/pods/964cd051-6afe-4834-a7b8-ccbe2c547f4e/volumes" Oct 10 09:38:03 crc kubenswrapper[4669]: I1010 09:38:03.808820 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddb14f43-b1a9-4ba3-b6db-734948a9c722" path="/var/lib/kubelet/pods/ddb14f43-b1a9-4ba3-b6db-734948a9c722/volumes" Oct 10 09:38:05 crc kubenswrapper[4669]: I1010 09:38:05.796067 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:38:05 crc kubenswrapper[4669]: E1010 09:38:05.796728 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:38:06 crc kubenswrapper[4669]: I1010 09:38:06.029779 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-99sxw"] Oct 10 09:38:06 crc kubenswrapper[4669]: I1010 09:38:06.038485 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-99sxw"] Oct 10 09:38:07 crc kubenswrapper[4669]: I1010 09:38:07.050879 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-29prl"] Oct 10 09:38:07 crc kubenswrapper[4669]: I1010 09:38:07.060232 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-29prl"] Oct 10 09:38:07 crc kubenswrapper[4669]: I1010 09:38:07.809270 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="680da54f-bdac-4014-8fd5-92f1cb5e715c" path="/var/lib/kubelet/pods/680da54f-bdac-4014-8fd5-92f1cb5e715c/volumes" Oct 10 09:38:07 crc kubenswrapper[4669]: I1010 09:38:07.810122 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53" path="/var/lib/kubelet/pods/ac28083f-b7bb-4f0e-aa1e-e1b69eaa8d53/volumes" Oct 10 09:38:09 crc kubenswrapper[4669]: I1010 09:38:09.840627 4669 scope.go:117] "RemoveContainer" containerID="c2b05e56b87acff4dc70c091af8c816c8013320e03be0c03e899054fbb519492" Oct 10 09:38:09 crc kubenswrapper[4669]: I1010 09:38:09.876607 4669 scope.go:117] "RemoveContainer" containerID="87652d4373bd9770f7bb009fd863ec138287023d6b394c7b0608446578e3c6df" Oct 10 09:38:09 crc kubenswrapper[4669]: I1010 09:38:09.936553 4669 scope.go:117] "RemoveContainer" containerID="09df8d6ef49cd88e2f1956c94859dbb324cade7f8699dfadd74f2104497986d5" Oct 10 09:38:09 crc kubenswrapper[4669]: I1010 09:38:09.967601 4669 scope.go:117] "RemoveContainer" containerID="f909e76202798459913f1cf80d080b20b34f1b6de27025771dd02fabdf311693" Oct 10 09:38:10 crc kubenswrapper[4669]: I1010 09:38:10.015167 4669 scope.go:117] "RemoveContainer" containerID="c38c3547c8ddcc5a50d07b3f2d76d69c047123990c58e1bb914f36b792f41196" Oct 10 09:38:10 crc kubenswrapper[4669]: I1010 09:38:10.056815 4669 scope.go:117] "RemoveContainer" containerID="f50c3b443c87455a1f988ec1ca512d2019dfb1a2e8e62c61039ca9547799fdb1" Oct 10 09:38:10 crc kubenswrapper[4669]: I1010 09:38:10.097202 4669 scope.go:117] "RemoveContainer" containerID="c418e67ac115abd378150868c2caee1b29bf9cfa73128fa8758b011bed59ad52" Oct 10 09:38:10 crc kubenswrapper[4669]: I1010 09:38:10.121124 4669 scope.go:117] "RemoveContainer" containerID="0e5a72e7d38c9c30f1b85be8f13eb21e397b8e22f18dcd0d590b4ed94b10a110" Oct 10 09:38:10 crc kubenswrapper[4669]: I1010 09:38:10.143388 4669 scope.go:117] "RemoveContainer" containerID="d2168a9083391f5a4c7edbf68ea3c7b9810c20f575156eca65bae1c65db4e44c" Oct 10 09:38:10 crc kubenswrapper[4669]: I1010 09:38:10.162111 4669 scope.go:117] "RemoveContainer" containerID="462f97283198a175384ec8ae0d73744ccb89d9a7c2532829e0512bc21036aaff" Oct 10 09:38:10 crc kubenswrapper[4669]: I1010 09:38:10.179529 4669 scope.go:117] "RemoveContainer" containerID="51543363f25e9768823f820b2bbaadd13707ccd3a1803e5b0408d0d5c003ecdc" Oct 10 09:38:16 crc kubenswrapper[4669]: I1010 09:38:16.795909 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:38:16 crc kubenswrapper[4669]: E1010 09:38:16.796945 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.049789 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5a0a-account-create-pbfzv"] Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.061897 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-af9a-account-create-hhvnc"] Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.070993 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-af9a-account-create-hhvnc"] Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.079094 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5a0a-account-create-pbfzv"] Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.090875 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-4851-account-create-lctwn"] Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.099713 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-4851-account-create-lctwn"] Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.264164 4669 generic.go:334] "Generic (PLEG): container finished" podID="f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" containerID="a9e6be0de9ee8269114d300be5825c5e168fd33df286bfb9dbe3d6b3dd594302" exitCode=0 Oct 10 09:38:20 crc kubenswrapper[4669]: I1010 09:38:20.264216 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" event={"ID":"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148","Type":"ContainerDied","Data":"a9e6be0de9ee8269114d300be5825c5e168fd33df286bfb9dbe3d6b3dd594302"} Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.674417 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.798023 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5ksg\" (UniqueName: \"kubernetes.io/projected/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-kube-api-access-l5ksg\") pod \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.798330 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-inventory\") pod \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.798381 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-ssh-key\") pod \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\" (UID: \"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148\") " Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.805777 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-kube-api-access-l5ksg" (OuterVolumeSpecName: "kube-api-access-l5ksg") pod "f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" (UID: "f4f5ee49-2ac8-49cf-b570-eaa3ad05c148"). InnerVolumeSpecName "kube-api-access-l5ksg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.807666 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00ebe613-c446-4805-800e-a2ce24c40bf5" path="/var/lib/kubelet/pods/00ebe613-c446-4805-800e-a2ce24c40bf5/volumes" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.808961 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf031e52-cb5b-49be-be4f-c136299774fe" path="/var/lib/kubelet/pods/bf031e52-cb5b-49be-be4f-c136299774fe/volumes" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.810185 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4512bc9-9144-4bbe-8ab4-7bdb096f18aa" path="/var/lib/kubelet/pods/d4512bc9-9144-4bbe-8ab4-7bdb096f18aa/volumes" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.827420 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" (UID: "f4f5ee49-2ac8-49cf-b570-eaa3ad05c148"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.827895 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-inventory" (OuterVolumeSpecName: "inventory") pod "f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" (UID: "f4f5ee49-2ac8-49cf-b570-eaa3ad05c148"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.900866 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5ksg\" (UniqueName: \"kubernetes.io/projected/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-kube-api-access-l5ksg\") on node \"crc\" DevicePath \"\"" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.901118 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:38:21 crc kubenswrapper[4669]: I1010 09:38:21.901148 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.032650 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-z8vqt"] Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.040724 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-z8vqt"] Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.301355 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.301508 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l" event={"ID":"f4f5ee49-2ac8-49cf-b570-eaa3ad05c148","Type":"ContainerDied","Data":"f0145a1a774b532e74d4b8fabc4461238b2d75c1cd40a2bd1a1fc28e775da258"} Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.301639 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0145a1a774b532e74d4b8fabc4461238b2d75c1cd40a2bd1a1fc28e775da258" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.367018 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz"] Oct 10 09:38:22 crc kubenswrapper[4669]: E1010 09:38:22.367757 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.367780 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.368008 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.368735 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.374457 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.374882 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.375043 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.377839 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.379731 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz"] Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.510729 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.510809 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.510894 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbqdl\" (UniqueName: \"kubernetes.io/projected/4505ce35-c314-4aef-b26f-821513df946f-kube-api-access-dbqdl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.612257 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.612329 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.612423 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbqdl\" (UniqueName: \"kubernetes.io/projected/4505ce35-c314-4aef-b26f-821513df946f-kube-api-access-dbqdl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.616698 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.617923 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.633200 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbqdl\" (UniqueName: \"kubernetes.io/projected/4505ce35-c314-4aef-b26f-821513df946f-kube-api-access-dbqdl\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:22 crc kubenswrapper[4669]: I1010 09:38:22.687219 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:23 crc kubenswrapper[4669]: I1010 09:38:23.230062 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz"] Oct 10 09:38:23 crc kubenswrapper[4669]: W1010 09:38:23.246366 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4505ce35_c314_4aef_b26f_821513df946f.slice/crio-8c3c75aad1a3c1a7bd9074e0a2180db3a0344a3a5f3cb9f0ad8add28b62a94ac WatchSource:0}: Error finding container 8c3c75aad1a3c1a7bd9074e0a2180db3a0344a3a5f3cb9f0ad8add28b62a94ac: Status 404 returned error can't find the container with id 8c3c75aad1a3c1a7bd9074e0a2180db3a0344a3a5f3cb9f0ad8add28b62a94ac Oct 10 09:38:23 crc kubenswrapper[4669]: I1010 09:38:23.313367 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" event={"ID":"4505ce35-c314-4aef-b26f-821513df946f","Type":"ContainerStarted","Data":"8c3c75aad1a3c1a7bd9074e0a2180db3a0344a3a5f3cb9f0ad8add28b62a94ac"} Oct 10 09:38:23 crc kubenswrapper[4669]: I1010 09:38:23.814340 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="799cf433-723c-4168-8522-773ca2947ec3" path="/var/lib/kubelet/pods/799cf433-723c-4168-8522-773ca2947ec3/volumes" Oct 10 09:38:24 crc kubenswrapper[4669]: I1010 09:38:24.325783 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" event={"ID":"4505ce35-c314-4aef-b26f-821513df946f","Type":"ContainerStarted","Data":"a11a8bd3394ecf53974839036494f0350151dcfca37e4dbc7cab550b63b8df1b"} Oct 10 09:38:24 crc kubenswrapper[4669]: I1010 09:38:24.347286 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" podStartSLOduration=1.522830951 podStartE2EDuration="2.347194761s" podCreationTimestamp="2025-10-10 09:38:22 +0000 UTC" firstStartedPulling="2025-10-10 09:38:23.250279736 +0000 UTC m=+1646.266298488" lastFinishedPulling="2025-10-10 09:38:24.074643516 +0000 UTC m=+1647.090662298" observedRunningTime="2025-10-10 09:38:24.338487342 +0000 UTC m=+1647.354506084" watchObservedRunningTime="2025-10-10 09:38:24.347194761 +0000 UTC m=+1647.363213503" Oct 10 09:38:26 crc kubenswrapper[4669]: I1010 09:38:26.038916 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-lm4k9"] Oct 10 09:38:26 crc kubenswrapper[4669]: I1010 09:38:26.048117 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-lm4k9"] Oct 10 09:38:27 crc kubenswrapper[4669]: I1010 09:38:27.809457 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1" path="/var/lib/kubelet/pods/6ea3c637-dcb5-4b7c-827f-f673f9d7d8d1/volumes" Oct 10 09:38:29 crc kubenswrapper[4669]: I1010 09:38:29.796122 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:38:29 crc kubenswrapper[4669]: E1010 09:38:29.796634 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:38:30 crc kubenswrapper[4669]: I1010 09:38:30.372423 4669 generic.go:334] "Generic (PLEG): container finished" podID="4505ce35-c314-4aef-b26f-821513df946f" containerID="a11a8bd3394ecf53974839036494f0350151dcfca37e4dbc7cab550b63b8df1b" exitCode=0 Oct 10 09:38:30 crc kubenswrapper[4669]: I1010 09:38:30.372489 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" event={"ID":"4505ce35-c314-4aef-b26f-821513df946f","Type":"ContainerDied","Data":"a11a8bd3394ecf53974839036494f0350151dcfca37e4dbc7cab550b63b8df1b"} Oct 10 09:38:31 crc kubenswrapper[4669]: I1010 09:38:31.832006 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:31 crc kubenswrapper[4669]: I1010 09:38:31.986705 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbqdl\" (UniqueName: \"kubernetes.io/projected/4505ce35-c314-4aef-b26f-821513df946f-kube-api-access-dbqdl\") pod \"4505ce35-c314-4aef-b26f-821513df946f\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " Oct 10 09:38:31 crc kubenswrapper[4669]: I1010 09:38:31.986825 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-inventory\") pod \"4505ce35-c314-4aef-b26f-821513df946f\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " Oct 10 09:38:31 crc kubenswrapper[4669]: I1010 09:38:31.986987 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-ssh-key\") pod \"4505ce35-c314-4aef-b26f-821513df946f\" (UID: \"4505ce35-c314-4aef-b26f-821513df946f\") " Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.000949 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4505ce35-c314-4aef-b26f-821513df946f-kube-api-access-dbqdl" (OuterVolumeSpecName: "kube-api-access-dbqdl") pod "4505ce35-c314-4aef-b26f-821513df946f" (UID: "4505ce35-c314-4aef-b26f-821513df946f"). InnerVolumeSpecName "kube-api-access-dbqdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.033093 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-inventory" (OuterVolumeSpecName: "inventory") pod "4505ce35-c314-4aef-b26f-821513df946f" (UID: "4505ce35-c314-4aef-b26f-821513df946f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.048785 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4505ce35-c314-4aef-b26f-821513df946f" (UID: "4505ce35-c314-4aef-b26f-821513df946f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.091568 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbqdl\" (UniqueName: \"kubernetes.io/projected/4505ce35-c314-4aef-b26f-821513df946f-kube-api-access-dbqdl\") on node \"crc\" DevicePath \"\"" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.099652 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.099868 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4505ce35-c314-4aef-b26f-821513df946f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.392914 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" event={"ID":"4505ce35-c314-4aef-b26f-821513df946f","Type":"ContainerDied","Data":"8c3c75aad1a3c1a7bd9074e0a2180db3a0344a3a5f3cb9f0ad8add28b62a94ac"} Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.393403 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c3c75aad1a3c1a7bd9074e0a2180db3a0344a3a5f3cb9f0ad8add28b62a94ac" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.393009 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.462241 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp"] Oct 10 09:38:32 crc kubenswrapper[4669]: E1010 09:38:32.462650 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4505ce35-c314-4aef-b26f-821513df946f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.462674 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="4505ce35-c314-4aef-b26f-821513df946f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.462841 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="4505ce35-c314-4aef-b26f-821513df946f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.463397 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.466180 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.466361 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.466545 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.475956 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.488345 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp"] Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.506131 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.506356 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.506496 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwdh7\" (UniqueName: \"kubernetes.io/projected/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-kube-api-access-mwdh7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.608553 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.608661 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwdh7\" (UniqueName: \"kubernetes.io/projected/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-kube-api-access-mwdh7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.608783 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.612316 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.613882 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.627124 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwdh7\" (UniqueName: \"kubernetes.io/projected/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-kube-api-access-mwdh7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-n2bvp\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:32 crc kubenswrapper[4669]: I1010 09:38:32.784326 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:38:33 crc kubenswrapper[4669]: I1010 09:38:33.123703 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp"] Oct 10 09:38:33 crc kubenswrapper[4669]: I1010 09:38:33.402434 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" event={"ID":"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57","Type":"ContainerStarted","Data":"8ecf91b63b38ef0a47b13b0449a1ba3320f73e31367821e80f779a515b3e4072"} Oct 10 09:38:34 crc kubenswrapper[4669]: I1010 09:38:34.415208 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" event={"ID":"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57","Type":"ContainerStarted","Data":"01ada31687f73fea7b0ccec2719af3a7f1b8022d7926549b3a45cfa0538a3253"} Oct 10 09:38:34 crc kubenswrapper[4669]: I1010 09:38:34.438127 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" podStartSLOduration=1.929360244 podStartE2EDuration="2.438109299s" podCreationTimestamp="2025-10-10 09:38:32 +0000 UTC" firstStartedPulling="2025-10-10 09:38:33.129012962 +0000 UTC m=+1656.145031714" lastFinishedPulling="2025-10-10 09:38:33.637762007 +0000 UTC m=+1656.653780769" observedRunningTime="2025-10-10 09:38:34.437205619 +0000 UTC m=+1657.453224371" watchObservedRunningTime="2025-10-10 09:38:34.438109299 +0000 UTC m=+1657.454128041" Oct 10 09:38:44 crc kubenswrapper[4669]: I1010 09:38:44.795384 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:38:44 crc kubenswrapper[4669]: E1010 09:38:44.796318 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:38:50 crc kubenswrapper[4669]: I1010 09:38:50.045138 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-l7kr8"] Oct 10 09:38:50 crc kubenswrapper[4669]: I1010 09:38:50.055183 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-l7kr8"] Oct 10 09:38:51 crc kubenswrapper[4669]: I1010 09:38:51.040174 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-d8m9m"] Oct 10 09:38:51 crc kubenswrapper[4669]: I1010 09:38:51.047720 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-d8m9m"] Oct 10 09:38:51 crc kubenswrapper[4669]: I1010 09:38:51.858656 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff" path="/var/lib/kubelet/pods/37ed6a10-c32a-4f82-8fa6-63e0cd8de8ff/volumes" Oct 10 09:38:51 crc kubenswrapper[4669]: I1010 09:38:51.859192 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50102622-b6c0-4fe9-8eab-b4027da1a36a" path="/var/lib/kubelet/pods/50102622-b6c0-4fe9-8eab-b4027da1a36a/volumes" Oct 10 09:38:59 crc kubenswrapper[4669]: I1010 09:38:59.795416 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:38:59 crc kubenswrapper[4669]: E1010 09:38:59.796847 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:39:10 crc kubenswrapper[4669]: I1010 09:39:10.447835 4669 scope.go:117] "RemoveContainer" containerID="19a34cf929c657db2e9a3c7f84851bfeae4662057f442aee63d4136bcc10fb15" Oct 10 09:39:10 crc kubenswrapper[4669]: I1010 09:39:10.478754 4669 scope.go:117] "RemoveContainer" containerID="5897be1a6b0542d0b527f840ebc2861fa37c672cec4e9488e2ae8dc51bfcb40a" Oct 10 09:39:10 crc kubenswrapper[4669]: I1010 09:39:10.517026 4669 scope.go:117] "RemoveContainer" containerID="10124d233567fa84e3a8db93abb7b4b87d2aee9f10df8b204cb7a79c0213aec0" Oct 10 09:39:10 crc kubenswrapper[4669]: I1010 09:39:10.558038 4669 scope.go:117] "RemoveContainer" containerID="ee1036c5e473647e3b0c63ba754ff95245c3d11d88512f8e661b954db1f591c0" Oct 10 09:39:10 crc kubenswrapper[4669]: I1010 09:39:10.608503 4669 scope.go:117] "RemoveContainer" containerID="416e34c6fab75954a71767e78290564c70e9e1feae973fab44258c9fa061165f" Oct 10 09:39:10 crc kubenswrapper[4669]: I1010 09:39:10.644757 4669 scope.go:117] "RemoveContainer" containerID="0eb010860136cb69a607f20a56d50bc6cb600cc3aa977818b62bfe45edf66223" Oct 10 09:39:10 crc kubenswrapper[4669]: I1010 09:39:10.692427 4669 scope.go:117] "RemoveContainer" containerID="d778f8e40a75b782c3fcb763fc57410d9ffc558284a07569abfe8a957d75a712" Oct 10 09:39:12 crc kubenswrapper[4669]: I1010 09:39:12.795510 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:39:12 crc kubenswrapper[4669]: E1010 09:39:12.795920 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:39:18 crc kubenswrapper[4669]: I1010 09:39:18.044051 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-trw89"] Oct 10 09:39:18 crc kubenswrapper[4669]: I1010 09:39:18.052299 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-trw89"] Oct 10 09:39:18 crc kubenswrapper[4669]: I1010 09:39:18.837064 4669 generic.go:334] "Generic (PLEG): container finished" podID="cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" containerID="01ada31687f73fea7b0ccec2719af3a7f1b8022d7926549b3a45cfa0538a3253" exitCode=0 Oct 10 09:39:18 crc kubenswrapper[4669]: I1010 09:39:18.837108 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" event={"ID":"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57","Type":"ContainerDied","Data":"01ada31687f73fea7b0ccec2719af3a7f1b8022d7926549b3a45cfa0538a3253"} Oct 10 09:39:19 crc kubenswrapper[4669]: I1010 09:39:19.817111 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1d9870d-304c-4ab1-9091-9e7e939cc6b8" path="/var/lib/kubelet/pods/a1d9870d-304c-4ab1-9091-9e7e939cc6b8/volumes" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.241441 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.345113 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-ssh-key\") pod \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.345224 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-inventory\") pod \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.345265 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwdh7\" (UniqueName: \"kubernetes.io/projected/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-kube-api-access-mwdh7\") pod \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\" (UID: \"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57\") " Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.351623 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-kube-api-access-mwdh7" (OuterVolumeSpecName: "kube-api-access-mwdh7") pod "cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" (UID: "cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57"). InnerVolumeSpecName "kube-api-access-mwdh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.373948 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" (UID: "cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.392770 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-inventory" (OuterVolumeSpecName: "inventory") pod "cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" (UID: "cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.447441 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.447489 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.447523 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwdh7\" (UniqueName: \"kubernetes.io/projected/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57-kube-api-access-mwdh7\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.859650 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" event={"ID":"cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57","Type":"ContainerDied","Data":"8ecf91b63b38ef0a47b13b0449a1ba3320f73e31367821e80f779a515b3e4072"} Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.859702 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ecf91b63b38ef0a47b13b0449a1ba3320f73e31367821e80f779a515b3e4072" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.859747 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.905173 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hsdq4"] Oct 10 09:39:20 crc kubenswrapper[4669]: E1010 09:39:20.905610 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.905635 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.905863 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.907415 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.917722 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsdq4"] Oct 10 09:39:20 crc kubenswrapper[4669]: I1010 09:39:20.999283 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh"] Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.000899 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.003751 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.004015 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.004124 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.004381 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.011709 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh"] Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.061930 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhz4b\" (UniqueName: \"kubernetes.io/projected/f95a20ee-12f1-4a65-8444-f8689da70525-kube-api-access-hhz4b\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.062205 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.062321 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7z6v\" (UniqueName: \"kubernetes.io/projected/bf129287-698a-41fd-8623-83628590e428-kube-api-access-m7z6v\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.062448 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.062665 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-catalog-content\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.062842 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-utilities\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.164044 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7z6v\" (UniqueName: \"kubernetes.io/projected/bf129287-698a-41fd-8623-83628590e428-kube-api-access-m7z6v\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.164461 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.164514 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-catalog-content\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.164567 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-utilities\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.164799 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhz4b\" (UniqueName: \"kubernetes.io/projected/f95a20ee-12f1-4a65-8444-f8689da70525-kube-api-access-hhz4b\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.164843 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.165131 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-catalog-content\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.165219 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-utilities\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.169898 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.182842 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.187415 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhz4b\" (UniqueName: \"kubernetes.io/projected/f95a20ee-12f1-4a65-8444-f8689da70525-kube-api-access-hhz4b\") pod \"certified-operators-hsdq4\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.193336 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7z6v\" (UniqueName: \"kubernetes.io/projected/bf129287-698a-41fd-8623-83628590e428-kube-api-access-m7z6v\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.235315 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.317272 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.633785 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsdq4"] Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.868853 4669 generic.go:334] "Generic (PLEG): container finished" podID="f95a20ee-12f1-4a65-8444-f8689da70525" containerID="a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db" exitCode=0 Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.868922 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdq4" event={"ID":"f95a20ee-12f1-4a65-8444-f8689da70525","Type":"ContainerDied","Data":"a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db"} Oct 10 09:39:21 crc kubenswrapper[4669]: I1010 09:39:21.868978 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdq4" event={"ID":"f95a20ee-12f1-4a65-8444-f8689da70525","Type":"ContainerStarted","Data":"8826e38e04b01e7e55f860a869ebaa3edef71874b515c80ef29f4c524f8d09d1"} Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.009501 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh"] Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.043163 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-qdk89"] Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.054204 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-zwrh7"] Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.061278 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-zwrh7"] Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.067506 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-qdk89"] Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.882993 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" event={"ID":"bf129287-698a-41fd-8623-83628590e428","Type":"ContainerStarted","Data":"66cb8d5f3b05585c190d8238874135c2ba38b9d5538df9755820b58eaed6165b"} Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.883445 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" event={"ID":"bf129287-698a-41fd-8623-83628590e428","Type":"ContainerStarted","Data":"401666bdbafd6072fbc5d1cf0068927c4c32128e0898367ebfa3a4ad36ef1bfe"} Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.889358 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdq4" event={"ID":"f95a20ee-12f1-4a65-8444-f8689da70525","Type":"ContainerStarted","Data":"7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c"} Oct 10 09:39:22 crc kubenswrapper[4669]: I1010 09:39:22.906010 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" podStartSLOduration=2.281510384 podStartE2EDuration="2.905992178s" podCreationTimestamp="2025-10-10 09:39:20 +0000 UTC" firstStartedPulling="2025-10-10 09:39:22.02645878 +0000 UTC m=+1705.042477532" lastFinishedPulling="2025-10-10 09:39:22.650940584 +0000 UTC m=+1705.666959326" observedRunningTime="2025-10-10 09:39:22.897565808 +0000 UTC m=+1705.913584550" watchObservedRunningTime="2025-10-10 09:39:22.905992178 +0000 UTC m=+1705.922010920" Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.033828 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-l66br"] Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.041295 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-l66br"] Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.796632 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:39:23 crc kubenswrapper[4669]: E1010 09:39:23.797540 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.816901 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="343f638d-3bd8-4737-99ee-a94844f9a261" path="/var/lib/kubelet/pods/343f638d-3bd8-4737-99ee-a94844f9a261/volumes" Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.818621 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4fe92a4-97c8-4787-94c3-89ae76827800" path="/var/lib/kubelet/pods/a4fe92a4-97c8-4787-94c3-89ae76827800/volumes" Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.819989 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6725ca7-4c3f-4e08-b053-d8eeeef15394" path="/var/lib/kubelet/pods/d6725ca7-4c3f-4e08-b053-d8eeeef15394/volumes" Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.922261 4669 generic.go:334] "Generic (PLEG): container finished" podID="f95a20ee-12f1-4a65-8444-f8689da70525" containerID="7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c" exitCode=0 Oct 10 09:39:23 crc kubenswrapper[4669]: I1010 09:39:23.922440 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdq4" event={"ID":"f95a20ee-12f1-4a65-8444-f8689da70525","Type":"ContainerDied","Data":"7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c"} Oct 10 09:39:24 crc kubenswrapper[4669]: I1010 09:39:24.932819 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdq4" event={"ID":"f95a20ee-12f1-4a65-8444-f8689da70525","Type":"ContainerStarted","Data":"60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c"} Oct 10 09:39:24 crc kubenswrapper[4669]: I1010 09:39:24.953923 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hsdq4" podStartSLOduration=2.44615824 podStartE2EDuration="4.953905962s" podCreationTimestamp="2025-10-10 09:39:20 +0000 UTC" firstStartedPulling="2025-10-10 09:39:21.870828032 +0000 UTC m=+1704.886846784" lastFinishedPulling="2025-10-10 09:39:24.378575754 +0000 UTC m=+1707.394594506" observedRunningTime="2025-10-10 09:39:24.949712468 +0000 UTC m=+1707.965731230" watchObservedRunningTime="2025-10-10 09:39:24.953905962 +0000 UTC m=+1707.969924704" Oct 10 09:39:26 crc kubenswrapper[4669]: I1010 09:39:26.947272 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf129287-698a-41fd-8623-83628590e428" containerID="66cb8d5f3b05585c190d8238874135c2ba38b9d5538df9755820b58eaed6165b" exitCode=0 Oct 10 09:39:26 crc kubenswrapper[4669]: I1010 09:39:26.947441 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" event={"ID":"bf129287-698a-41fd-8623-83628590e428","Type":"ContainerDied","Data":"66cb8d5f3b05585c190d8238874135c2ba38b9d5538df9755820b58eaed6165b"} Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.395435 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.436590 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-inventory\") pod \"bf129287-698a-41fd-8623-83628590e428\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.436732 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7z6v\" (UniqueName: \"kubernetes.io/projected/bf129287-698a-41fd-8623-83628590e428-kube-api-access-m7z6v\") pod \"bf129287-698a-41fd-8623-83628590e428\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.436885 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-ssh-key\") pod \"bf129287-698a-41fd-8623-83628590e428\" (UID: \"bf129287-698a-41fd-8623-83628590e428\") " Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.448161 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf129287-698a-41fd-8623-83628590e428-kube-api-access-m7z6v" (OuterVolumeSpecName: "kube-api-access-m7z6v") pod "bf129287-698a-41fd-8623-83628590e428" (UID: "bf129287-698a-41fd-8623-83628590e428"). InnerVolumeSpecName "kube-api-access-m7z6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.466156 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf129287-698a-41fd-8623-83628590e428" (UID: "bf129287-698a-41fd-8623-83628590e428"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.473175 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-inventory" (OuterVolumeSpecName: "inventory") pod "bf129287-698a-41fd-8623-83628590e428" (UID: "bf129287-698a-41fd-8623-83628590e428"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.538577 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.538632 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7z6v\" (UniqueName: \"kubernetes.io/projected/bf129287-698a-41fd-8623-83628590e428-kube-api-access-m7z6v\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.538646 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf129287-698a-41fd-8623-83628590e428-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.972497 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" event={"ID":"bf129287-698a-41fd-8623-83628590e428","Type":"ContainerDied","Data":"401666bdbafd6072fbc5d1cf0068927c4c32128e0898367ebfa3a4ad36ef1bfe"} Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.972540 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="401666bdbafd6072fbc5d1cf0068927c4c32128e0898367ebfa3a4ad36ef1bfe" Oct 10 09:39:28 crc kubenswrapper[4669]: I1010 09:39:28.972629 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.065064 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67"] Oct 10 09:39:29 crc kubenswrapper[4669]: E1010 09:39:29.065707 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf129287-698a-41fd-8623-83628590e428" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.065738 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf129287-698a-41fd-8623-83628590e428" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.066090 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf129287-698a-41fd-8623-83628590e428" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.067115 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.069915 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.070252 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.071430 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.071726 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.074001 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67"] Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.151354 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.151419 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.151465 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pzd4\" (UniqueName: \"kubernetes.io/projected/8c21d000-f125-4187-bafc-75c372f943a1-kube-api-access-2pzd4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.252823 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.252888 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pzd4\" (UniqueName: \"kubernetes.io/projected/8c21d000-f125-4187-bafc-75c372f943a1-kube-api-access-2pzd4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.253056 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.256762 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.266057 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.269803 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pzd4\" (UniqueName: \"kubernetes.io/projected/8c21d000-f125-4187-bafc-75c372f943a1-kube-api-access-2pzd4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-g5w67\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.419053 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.955005 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67"] Oct 10 09:39:29 crc kubenswrapper[4669]: I1010 09:39:29.984465 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" event={"ID":"8c21d000-f125-4187-bafc-75c372f943a1","Type":"ContainerStarted","Data":"80e52eaa787098bc00122f0c8b6de7c0e87404b79f0ff06f60b39ad8f97ffca6"} Oct 10 09:39:31 crc kubenswrapper[4669]: I1010 09:39:31.236552 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:31 crc kubenswrapper[4669]: I1010 09:39:31.236870 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:31 crc kubenswrapper[4669]: I1010 09:39:31.286220 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:32 crc kubenswrapper[4669]: I1010 09:39:32.007732 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" event={"ID":"8c21d000-f125-4187-bafc-75c372f943a1","Type":"ContainerStarted","Data":"25d9022310ee71ba3703d87ff416aa4d5a3ff840a1e9f29959aa518c4ddde469"} Oct 10 09:39:32 crc kubenswrapper[4669]: I1010 09:39:32.035816 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" podStartSLOduration=2.192723002 podStartE2EDuration="3.035794182s" podCreationTimestamp="2025-10-10 09:39:29 +0000 UTC" firstStartedPulling="2025-10-10 09:39:29.964136097 +0000 UTC m=+1712.980154849" lastFinishedPulling="2025-10-10 09:39:30.807207287 +0000 UTC m=+1713.823226029" observedRunningTime="2025-10-10 09:39:32.035700939 +0000 UTC m=+1715.051719691" watchObservedRunningTime="2025-10-10 09:39:32.035794182 +0000 UTC m=+1715.051812934" Oct 10 09:39:32 crc kubenswrapper[4669]: I1010 09:39:32.077164 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:32 crc kubenswrapper[4669]: I1010 09:39:32.143808 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hsdq4"] Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.028531 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hsdq4" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="registry-server" containerID="cri-o://60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c" gracePeriod=2 Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.457239 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.557400 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-utilities\") pod \"f95a20ee-12f1-4a65-8444-f8689da70525\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.557488 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhz4b\" (UniqueName: \"kubernetes.io/projected/f95a20ee-12f1-4a65-8444-f8689da70525-kube-api-access-hhz4b\") pod \"f95a20ee-12f1-4a65-8444-f8689da70525\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.557795 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-catalog-content\") pod \"f95a20ee-12f1-4a65-8444-f8689da70525\" (UID: \"f95a20ee-12f1-4a65-8444-f8689da70525\") " Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.558281 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-utilities" (OuterVolumeSpecName: "utilities") pod "f95a20ee-12f1-4a65-8444-f8689da70525" (UID: "f95a20ee-12f1-4a65-8444-f8689da70525"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.578538 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f95a20ee-12f1-4a65-8444-f8689da70525-kube-api-access-hhz4b" (OuterVolumeSpecName: "kube-api-access-hhz4b") pod "f95a20ee-12f1-4a65-8444-f8689da70525" (UID: "f95a20ee-12f1-4a65-8444-f8689da70525"). InnerVolumeSpecName "kube-api-access-hhz4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.605821 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f95a20ee-12f1-4a65-8444-f8689da70525" (UID: "f95a20ee-12f1-4a65-8444-f8689da70525"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.661750 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.661815 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f95a20ee-12f1-4a65-8444-f8689da70525-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:34 crc kubenswrapper[4669]: I1010 09:39:34.661836 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhz4b\" (UniqueName: \"kubernetes.io/projected/f95a20ee-12f1-4a65-8444-f8689da70525-kube-api-access-hhz4b\") on node \"crc\" DevicePath \"\"" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.045811 4669 generic.go:334] "Generic (PLEG): container finished" podID="f95a20ee-12f1-4a65-8444-f8689da70525" containerID="60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c" exitCode=0 Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.045898 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdq4" event={"ID":"f95a20ee-12f1-4a65-8444-f8689da70525","Type":"ContainerDied","Data":"60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c"} Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.045943 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsdq4" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.045986 4669 scope.go:117] "RemoveContainer" containerID="60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.045963 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsdq4" event={"ID":"f95a20ee-12f1-4a65-8444-f8689da70525","Type":"ContainerDied","Data":"8826e38e04b01e7e55f860a869ebaa3edef71874b515c80ef29f4c524f8d09d1"} Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.082147 4669 scope.go:117] "RemoveContainer" containerID="7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.124657 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hsdq4"] Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.131142 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hsdq4"] Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.135435 4669 scope.go:117] "RemoveContainer" containerID="a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.168743 4669 scope.go:117] "RemoveContainer" containerID="60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c" Oct 10 09:39:35 crc kubenswrapper[4669]: E1010 09:39:35.169518 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c\": container with ID starting with 60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c not found: ID does not exist" containerID="60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.169620 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c"} err="failed to get container status \"60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c\": rpc error: code = NotFound desc = could not find container \"60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c\": container with ID starting with 60d2ff3af1b905bf25d92c2520fc9043c8b2c95e80bf9976e6d4169bfb0f4d0c not found: ID does not exist" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.169698 4669 scope.go:117] "RemoveContainer" containerID="7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c" Oct 10 09:39:35 crc kubenswrapper[4669]: E1010 09:39:35.170121 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c\": container with ID starting with 7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c not found: ID does not exist" containerID="7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.170164 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c"} err="failed to get container status \"7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c\": rpc error: code = NotFound desc = could not find container \"7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c\": container with ID starting with 7cbd07036b9630116c14b1d539c828f9f9e7c7a77672a628d5651c579a42f68c not found: ID does not exist" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.170192 4669 scope.go:117] "RemoveContainer" containerID="a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db" Oct 10 09:39:35 crc kubenswrapper[4669]: E1010 09:39:35.170574 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db\": container with ID starting with a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db not found: ID does not exist" containerID="a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.170614 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db"} err="failed to get container status \"a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db\": rpc error: code = NotFound desc = could not find container \"a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db\": container with ID starting with a8a4144e405ba87a120723976a5596756d328b4c46ce4722410c5f6f277553db not found: ID does not exist" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.798396 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:39:35 crc kubenswrapper[4669]: E1010 09:39:35.798709 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:39:35 crc kubenswrapper[4669]: I1010 09:39:35.806285 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" path="/var/lib/kubelet/pods/f95a20ee-12f1-4a65-8444-f8689da70525/volumes" Oct 10 09:39:38 crc kubenswrapper[4669]: I1010 09:39:38.038740 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a005-account-create-drmrx"] Oct 10 09:39:38 crc kubenswrapper[4669]: I1010 09:39:38.046790 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-d5de-account-create-mxqm2"] Oct 10 09:39:38 crc kubenswrapper[4669]: I1010 09:39:38.056317 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c235-account-create-qclhl"] Oct 10 09:39:38 crc kubenswrapper[4669]: I1010 09:39:38.063433 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a005-account-create-drmrx"] Oct 10 09:39:38 crc kubenswrapper[4669]: I1010 09:39:38.070243 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-d5de-account-create-mxqm2"] Oct 10 09:39:38 crc kubenswrapper[4669]: I1010 09:39:38.076803 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c235-account-create-qclhl"] Oct 10 09:39:39 crc kubenswrapper[4669]: I1010 09:39:39.822250 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02fe8a8b-bb2b-4d50-b645-49f4fd745dbd" path="/var/lib/kubelet/pods/02fe8a8b-bb2b-4d50-b645-49f4fd745dbd/volumes" Oct 10 09:39:39 crc kubenswrapper[4669]: I1010 09:39:39.824133 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63afb443-38e7-4f99-98af-007b85175785" path="/var/lib/kubelet/pods/63afb443-38e7-4f99-98af-007b85175785/volumes" Oct 10 09:39:39 crc kubenswrapper[4669]: I1010 09:39:39.825661 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbb4d91d-a933-4c7f-bcf9-beab62822018" path="/var/lib/kubelet/pods/dbb4d91d-a933-4c7f-bcf9-beab62822018/volumes" Oct 10 09:39:49 crc kubenswrapper[4669]: I1010 09:39:49.796249 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:39:49 crc kubenswrapper[4669]: E1010 09:39:49.797512 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:40:04 crc kubenswrapper[4669]: I1010 09:40:04.795425 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:40:04 crc kubenswrapper[4669]: E1010 09:40:04.796201 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:40:08 crc kubenswrapper[4669]: I1010 09:40:08.060519 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-c7jmw"] Oct 10 09:40:08 crc kubenswrapper[4669]: I1010 09:40:08.071159 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-c7jmw"] Oct 10 09:40:09 crc kubenswrapper[4669]: I1010 09:40:09.806888 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12148914-1d9a-49a1-a080-e96fce7f1799" path="/var/lib/kubelet/pods/12148914-1d9a-49a1-a080-e96fce7f1799/volumes" Oct 10 09:40:10 crc kubenswrapper[4669]: I1010 09:40:10.836900 4669 scope.go:117] "RemoveContainer" containerID="4e6e1e2d9fa7339fbcfd04e37c884a1eb4d3cbfc91ed126a8acf41e26a14a3dc" Oct 10 09:40:10 crc kubenswrapper[4669]: I1010 09:40:10.882968 4669 scope.go:117] "RemoveContainer" containerID="8d0bf3c8764456a6044302d994777eef5bda57efc173b6dfed30d103f8783e0e" Oct 10 09:40:10 crc kubenswrapper[4669]: I1010 09:40:10.932462 4669 scope.go:117] "RemoveContainer" containerID="ea9573f19a0a7de26d8deddf86a4e6be88b99a8f4fcada8865d24c03d0ad7c55" Oct 10 09:40:10 crc kubenswrapper[4669]: I1010 09:40:10.953659 4669 scope.go:117] "RemoveContainer" containerID="218bc52fe18fe937533de4bc76efd2d255c6b498fa98167ebc7c538bf8d8bb3b" Oct 10 09:40:10 crc kubenswrapper[4669]: I1010 09:40:10.991185 4669 scope.go:117] "RemoveContainer" containerID="763bc4a1878d058c147ff64408d44018e9014c82686cac8366fafd168f150979" Oct 10 09:40:11 crc kubenswrapper[4669]: I1010 09:40:11.024422 4669 scope.go:117] "RemoveContainer" containerID="4ecebc4fe8bbce718d521be98fb4e54b676f457f449b6b5f66b255d9db6bce04" Oct 10 09:40:11 crc kubenswrapper[4669]: I1010 09:40:11.062974 4669 scope.go:117] "RemoveContainer" containerID="0f0a4e06557b0dbdd9cc9514819fe08cc24eff230edfea029fd9994358e39be6" Oct 10 09:40:11 crc kubenswrapper[4669]: I1010 09:40:11.080839 4669 scope.go:117] "RemoveContainer" containerID="24413fed3d637c1706004ffba8e7efda201d191ac13ca1af792d3082b5505611" Oct 10 09:40:15 crc kubenswrapper[4669]: I1010 09:40:15.796154 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:40:15 crc kubenswrapper[4669]: E1010 09:40:15.797229 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:40:26 crc kubenswrapper[4669]: I1010 09:40:26.797265 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:40:26 crc kubenswrapper[4669]: E1010 09:40:26.798159 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:40:29 crc kubenswrapper[4669]: I1010 09:40:29.050012 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8c4xv"] Oct 10 09:40:29 crc kubenswrapper[4669]: I1010 09:40:29.056860 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-6fl4m"] Oct 10 09:40:29 crc kubenswrapper[4669]: I1010 09:40:29.063644 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8c4xv"] Oct 10 09:40:29 crc kubenswrapper[4669]: I1010 09:40:29.070702 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-6fl4m"] Oct 10 09:40:29 crc kubenswrapper[4669]: I1010 09:40:29.807875 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="789ad8bf-d401-4d16-9752-9ffe5f153fc5" path="/var/lib/kubelet/pods/789ad8bf-d401-4d16-9752-9ffe5f153fc5/volumes" Oct 10 09:40:29 crc kubenswrapper[4669]: I1010 09:40:29.808642 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c1db2af-8211-4298-8514-047aa3e098cc" path="/var/lib/kubelet/pods/9c1db2af-8211-4298-8514-047aa3e098cc/volumes" Oct 10 09:40:32 crc kubenswrapper[4669]: I1010 09:40:32.581706 4669 generic.go:334] "Generic (PLEG): container finished" podID="8c21d000-f125-4187-bafc-75c372f943a1" containerID="25d9022310ee71ba3703d87ff416aa4d5a3ff840a1e9f29959aa518c4ddde469" exitCode=2 Oct 10 09:40:32 crc kubenswrapper[4669]: I1010 09:40:32.581792 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" event={"ID":"8c21d000-f125-4187-bafc-75c372f943a1","Type":"ContainerDied","Data":"25d9022310ee71ba3703d87ff416aa4d5a3ff840a1e9f29959aa518c4ddde469"} Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.031092 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.154942 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-inventory\") pod \"8c21d000-f125-4187-bafc-75c372f943a1\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.155031 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-ssh-key\") pod \"8c21d000-f125-4187-bafc-75c372f943a1\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.155127 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pzd4\" (UniqueName: \"kubernetes.io/projected/8c21d000-f125-4187-bafc-75c372f943a1-kube-api-access-2pzd4\") pod \"8c21d000-f125-4187-bafc-75c372f943a1\" (UID: \"8c21d000-f125-4187-bafc-75c372f943a1\") " Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.164743 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c21d000-f125-4187-bafc-75c372f943a1-kube-api-access-2pzd4" (OuterVolumeSpecName: "kube-api-access-2pzd4") pod "8c21d000-f125-4187-bafc-75c372f943a1" (UID: "8c21d000-f125-4187-bafc-75c372f943a1"). InnerVolumeSpecName "kube-api-access-2pzd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.187277 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8c21d000-f125-4187-bafc-75c372f943a1" (UID: "8c21d000-f125-4187-bafc-75c372f943a1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.194089 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-inventory" (OuterVolumeSpecName: "inventory") pod "8c21d000-f125-4187-bafc-75c372f943a1" (UID: "8c21d000-f125-4187-bafc-75c372f943a1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.257290 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pzd4\" (UniqueName: \"kubernetes.io/projected/8c21d000-f125-4187-bafc-75c372f943a1-kube-api-access-2pzd4\") on node \"crc\" DevicePath \"\"" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.257529 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.257702 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8c21d000-f125-4187-bafc-75c372f943a1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.600559 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" event={"ID":"8c21d000-f125-4187-bafc-75c372f943a1","Type":"ContainerDied","Data":"80e52eaa787098bc00122f0c8b6de7c0e87404b79f0ff06f60b39ad8f97ffca6"} Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.600615 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80e52eaa787098bc00122f0c8b6de7c0e87404b79f0ff06f60b39ad8f97ffca6" Oct 10 09:40:34 crc kubenswrapper[4669]: I1010 09:40:34.600623 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67" Oct 10 09:40:39 crc kubenswrapper[4669]: I1010 09:40:39.800536 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:40:39 crc kubenswrapper[4669]: E1010 09:40:39.801997 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.047748 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv"] Oct 10 09:40:41 crc kubenswrapper[4669]: E1010 09:40:41.048566 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="extract-content" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.048640 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="extract-content" Oct 10 09:40:41 crc kubenswrapper[4669]: E1010 09:40:41.048681 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="extract-utilities" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.048701 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="extract-utilities" Oct 10 09:40:41 crc kubenswrapper[4669]: E1010 09:40:41.048746 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="registry-server" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.048764 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="registry-server" Oct 10 09:40:41 crc kubenswrapper[4669]: E1010 09:40:41.048816 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c21d000-f125-4187-bafc-75c372f943a1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.048836 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c21d000-f125-4187-bafc-75c372f943a1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.049268 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f95a20ee-12f1-4a65-8444-f8689da70525" containerName="registry-server" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.049307 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c21d000-f125-4187-bafc-75c372f943a1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.050950 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.055384 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.055848 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.057158 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.060260 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.078075 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv"] Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.098881 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l6kr\" (UniqueName: \"kubernetes.io/projected/b814d616-5da3-4efb-844f-854a6ce7d878-kube-api-access-5l6kr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.098978 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.099070 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.201378 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l6kr\" (UniqueName: \"kubernetes.io/projected/b814d616-5da3-4efb-844f-854a6ce7d878-kube-api-access-5l6kr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.201441 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.201484 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.222700 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.224217 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l6kr\" (UniqueName: \"kubernetes.io/projected/b814d616-5da3-4efb-844f-854a6ce7d878-kube-api-access-5l6kr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.225052 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.398449 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:40:41 crc kubenswrapper[4669]: I1010 09:40:41.955907 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv"] Oct 10 09:40:42 crc kubenswrapper[4669]: I1010 09:40:42.685563 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" event={"ID":"b814d616-5da3-4efb-844f-854a6ce7d878","Type":"ContainerStarted","Data":"fdbddf72b681ed6299a502f8f4107c84bbfb589a05aaa20c78bf93193d4fcf0d"} Oct 10 09:40:43 crc kubenswrapper[4669]: I1010 09:40:43.697038 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" event={"ID":"b814d616-5da3-4efb-844f-854a6ce7d878","Type":"ContainerStarted","Data":"0df982b51daa067ddc60b0449155ca31fc67fafc8f809a0cac7d28c14734cbda"} Oct 10 09:40:43 crc kubenswrapper[4669]: I1010 09:40:43.723913 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" podStartSLOduration=2.128394549 podStartE2EDuration="2.723889533s" podCreationTimestamp="2025-10-10 09:40:41 +0000 UTC" firstStartedPulling="2025-10-10 09:40:41.958492914 +0000 UTC m=+1784.974511656" lastFinishedPulling="2025-10-10 09:40:42.553987888 +0000 UTC m=+1785.570006640" observedRunningTime="2025-10-10 09:40:43.71757279 +0000 UTC m=+1786.733591532" watchObservedRunningTime="2025-10-10 09:40:43.723889533 +0000 UTC m=+1786.739908275" Oct 10 09:40:52 crc kubenswrapper[4669]: I1010 09:40:52.796347 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:40:52 crc kubenswrapper[4669]: E1010 09:40:52.797327 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:41:06 crc kubenswrapper[4669]: I1010 09:41:06.796456 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:41:06 crc kubenswrapper[4669]: E1010 09:41:06.797273 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:41:11 crc kubenswrapper[4669]: I1010 09:41:11.056132 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-t6zc6"] Oct 10 09:41:11 crc kubenswrapper[4669]: I1010 09:41:11.069731 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-t6zc6"] Oct 10 09:41:11 crc kubenswrapper[4669]: I1010 09:41:11.254599 4669 scope.go:117] "RemoveContainer" containerID="8ed168dfddb073768e6fcdebc0da54e8d6dba95243a136a316671bd552c9b433" Oct 10 09:41:11 crc kubenswrapper[4669]: I1010 09:41:11.307214 4669 scope.go:117] "RemoveContainer" containerID="4aba981733fe9c39243143c7dfad566acf8c7b548a7a480acbaa1632d50a284b" Oct 10 09:41:11 crc kubenswrapper[4669]: I1010 09:41:11.809979 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc4bf0c-125d-4c30-9389-22b7b80b56f2" path="/var/lib/kubelet/pods/7fc4bf0c-125d-4c30-9389-22b7b80b56f2/volumes" Oct 10 09:41:18 crc kubenswrapper[4669]: I1010 09:41:18.795038 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:41:18 crc kubenswrapper[4669]: E1010 09:41:18.795905 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:41:30 crc kubenswrapper[4669]: I1010 09:41:30.795529 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:41:31 crc kubenswrapper[4669]: I1010 09:41:31.134601 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"3c8ef3ad011601a42242c3a0f4bbb5ca2056df6593c3e51f7cd342be4e813559"} Oct 10 09:41:36 crc kubenswrapper[4669]: I1010 09:41:36.195424 4669 generic.go:334] "Generic (PLEG): container finished" podID="b814d616-5da3-4efb-844f-854a6ce7d878" containerID="0df982b51daa067ddc60b0449155ca31fc67fafc8f809a0cac7d28c14734cbda" exitCode=0 Oct 10 09:41:36 crc kubenswrapper[4669]: I1010 09:41:36.195549 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" event={"ID":"b814d616-5da3-4efb-844f-854a6ce7d878","Type":"ContainerDied","Data":"0df982b51daa067ddc60b0449155ca31fc67fafc8f809a0cac7d28c14734cbda"} Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.675145 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.749445 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l6kr\" (UniqueName: \"kubernetes.io/projected/b814d616-5da3-4efb-844f-854a6ce7d878-kube-api-access-5l6kr\") pod \"b814d616-5da3-4efb-844f-854a6ce7d878\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.749618 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-ssh-key\") pod \"b814d616-5da3-4efb-844f-854a6ce7d878\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.749728 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-inventory\") pod \"b814d616-5da3-4efb-844f-854a6ce7d878\" (UID: \"b814d616-5da3-4efb-844f-854a6ce7d878\") " Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.755307 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b814d616-5da3-4efb-844f-854a6ce7d878-kube-api-access-5l6kr" (OuterVolumeSpecName: "kube-api-access-5l6kr") pod "b814d616-5da3-4efb-844f-854a6ce7d878" (UID: "b814d616-5da3-4efb-844f-854a6ce7d878"). InnerVolumeSpecName "kube-api-access-5l6kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.784834 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b814d616-5da3-4efb-844f-854a6ce7d878" (UID: "b814d616-5da3-4efb-844f-854a6ce7d878"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.785785 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-inventory" (OuterVolumeSpecName: "inventory") pod "b814d616-5da3-4efb-844f-854a6ce7d878" (UID: "b814d616-5da3-4efb-844f-854a6ce7d878"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.852656 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l6kr\" (UniqueName: \"kubernetes.io/projected/b814d616-5da3-4efb-844f-854a6ce7d878-kube-api-access-5l6kr\") on node \"crc\" DevicePath \"\"" Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.852737 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:41:37 crc kubenswrapper[4669]: I1010 09:41:37.852759 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b814d616-5da3-4efb-844f-854a6ce7d878-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.221711 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" event={"ID":"b814d616-5da3-4efb-844f-854a6ce7d878","Type":"ContainerDied","Data":"fdbddf72b681ed6299a502f8f4107c84bbfb589a05aaa20c78bf93193d4fcf0d"} Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.222414 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdbddf72b681ed6299a502f8f4107c84bbfb589a05aaa20c78bf93193d4fcf0d" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.221971 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.339240 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb99"] Oct 10 09:41:38 crc kubenswrapper[4669]: E1010 09:41:38.339564 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b814d616-5da3-4efb-844f-854a6ce7d878" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.339579 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b814d616-5da3-4efb-844f-854a6ce7d878" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.339809 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="b814d616-5da3-4efb-844f-854a6ce7d878" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.340336 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.351921 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.352005 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.352035 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.387077 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.419462 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb99"] Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.492837 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.492953 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.493065 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8wct\" (UniqueName: \"kubernetes.io/projected/1934fe86-0d23-4abb-be37-03beb45bd95d-kube-api-access-l8wct\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.594886 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8wct\" (UniqueName: \"kubernetes.io/projected/1934fe86-0d23-4abb-be37-03beb45bd95d-kube-api-access-l8wct\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.595190 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.595305 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.598478 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.598839 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.622310 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8wct\" (UniqueName: \"kubernetes.io/projected/1934fe86-0d23-4abb-be37-03beb45bd95d-kube-api-access-l8wct\") pod \"ssh-known-hosts-edpm-deployment-cmb99\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:38 crc kubenswrapper[4669]: I1010 09:41:38.668341 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:39 crc kubenswrapper[4669]: I1010 09:41:39.269404 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb99"] Oct 10 09:41:40 crc kubenswrapper[4669]: I1010 09:41:40.256215 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" event={"ID":"1934fe86-0d23-4abb-be37-03beb45bd95d","Type":"ContainerStarted","Data":"52c405f6cd5452db64620bb6f5cc4ec390feb43bb43501685970752ac0f0bb82"} Oct 10 09:41:40 crc kubenswrapper[4669]: I1010 09:41:40.256838 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" event={"ID":"1934fe86-0d23-4abb-be37-03beb45bd95d","Type":"ContainerStarted","Data":"0db4dbed92e9e5121f57a61de335a8031ed43bd61212b38b9ba46d4a26e3b0bb"} Oct 10 09:41:40 crc kubenswrapper[4669]: I1010 09:41:40.275403 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" podStartSLOduration=1.691147414 podStartE2EDuration="2.275388408s" podCreationTimestamp="2025-10-10 09:41:38 +0000 UTC" firstStartedPulling="2025-10-10 09:41:39.26202763 +0000 UTC m=+1842.278046412" lastFinishedPulling="2025-10-10 09:41:39.846268654 +0000 UTC m=+1842.862287406" observedRunningTime="2025-10-10 09:41:40.272728733 +0000 UTC m=+1843.288747475" watchObservedRunningTime="2025-10-10 09:41:40.275388408 +0000 UTC m=+1843.291407150" Oct 10 09:41:48 crc kubenswrapper[4669]: I1010 09:41:48.328054 4669 generic.go:334] "Generic (PLEG): container finished" podID="1934fe86-0d23-4abb-be37-03beb45bd95d" containerID="52c405f6cd5452db64620bb6f5cc4ec390feb43bb43501685970752ac0f0bb82" exitCode=0 Oct 10 09:41:48 crc kubenswrapper[4669]: I1010 09:41:48.328121 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" event={"ID":"1934fe86-0d23-4abb-be37-03beb45bd95d","Type":"ContainerDied","Data":"52c405f6cd5452db64620bb6f5cc4ec390feb43bb43501685970752ac0f0bb82"} Oct 10 09:41:49 crc kubenswrapper[4669]: I1010 09:41:49.807176 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:49 crc kubenswrapper[4669]: I1010 09:41:49.912627 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-ssh-key-openstack-edpm-ipam\") pod \"1934fe86-0d23-4abb-be37-03beb45bd95d\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " Oct 10 09:41:49 crc kubenswrapper[4669]: I1010 09:41:49.912963 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0\") pod \"1934fe86-0d23-4abb-be37-03beb45bd95d\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " Oct 10 09:41:49 crc kubenswrapper[4669]: I1010 09:41:49.913005 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8wct\" (UniqueName: \"kubernetes.io/projected/1934fe86-0d23-4abb-be37-03beb45bd95d-kube-api-access-l8wct\") pod \"1934fe86-0d23-4abb-be37-03beb45bd95d\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " Oct 10 09:41:49 crc kubenswrapper[4669]: I1010 09:41:49.918369 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1934fe86-0d23-4abb-be37-03beb45bd95d-kube-api-access-l8wct" (OuterVolumeSpecName: "kube-api-access-l8wct") pod "1934fe86-0d23-4abb-be37-03beb45bd95d" (UID: "1934fe86-0d23-4abb-be37-03beb45bd95d"). InnerVolumeSpecName "kube-api-access-l8wct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:41:49 crc kubenswrapper[4669]: E1010 09:41:49.940979 4669 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0 podName:1934fe86-0d23-4abb-be37-03beb45bd95d nodeName:}" failed. No retries permitted until 2025-10-10 09:41:50.440951061 +0000 UTC m=+1853.456969803 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory-0" (UniqueName: "kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0") pod "1934fe86-0d23-4abb-be37-03beb45bd95d" (UID: "1934fe86-0d23-4abb-be37-03beb45bd95d") : error deleting /var/lib/kubelet/pods/1934fe86-0d23-4abb-be37-03beb45bd95d/volume-subpaths: remove /var/lib/kubelet/pods/1934fe86-0d23-4abb-be37-03beb45bd95d/volume-subpaths: no such file or directory Oct 10 09:41:49 crc kubenswrapper[4669]: I1010 09:41:49.944114 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1934fe86-0d23-4abb-be37-03beb45bd95d" (UID: "1934fe86-0d23-4abb-be37-03beb45bd95d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.015131 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8wct\" (UniqueName: \"kubernetes.io/projected/1934fe86-0d23-4abb-be37-03beb45bd95d-kube-api-access-l8wct\") on node \"crc\" DevicePath \"\"" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.015175 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.351841 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" event={"ID":"1934fe86-0d23-4abb-be37-03beb45bd95d","Type":"ContainerDied","Data":"0db4dbed92e9e5121f57a61de335a8031ed43bd61212b38b9ba46d4a26e3b0bb"} Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.351914 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0db4dbed92e9e5121f57a61de335a8031ed43bd61212b38b9ba46d4a26e3b0bb" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.352014 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cmb99" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.468180 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx"] Oct 10 09:41:50 crc kubenswrapper[4669]: E1010 09:41:50.468553 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1934fe86-0d23-4abb-be37-03beb45bd95d" containerName="ssh-known-hosts-edpm-deployment" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.468564 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1934fe86-0d23-4abb-be37-03beb45bd95d" containerName="ssh-known-hosts-edpm-deployment" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.468802 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="1934fe86-0d23-4abb-be37-03beb45bd95d" containerName="ssh-known-hosts-edpm-deployment" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.469377 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.497347 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx"] Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.525285 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0\") pod \"1934fe86-0d23-4abb-be37-03beb45bd95d\" (UID: \"1934fe86-0d23-4abb-be37-03beb45bd95d\") " Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.528829 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "1934fe86-0d23-4abb-be37-03beb45bd95d" (UID: "1934fe86-0d23-4abb-be37-03beb45bd95d"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.629064 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spz29\" (UniqueName: \"kubernetes.io/projected/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-kube-api-access-spz29\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.629257 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.629599 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.629741 4669 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1934fe86-0d23-4abb-be37-03beb45bd95d-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.745442 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spz29\" (UniqueName: \"kubernetes.io/projected/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-kube-api-access-spz29\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.745828 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.745903 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.751174 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.751766 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.770944 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spz29\" (UniqueName: \"kubernetes.io/projected/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-kube-api-access-spz29\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mmkmx\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:50 crc kubenswrapper[4669]: I1010 09:41:50.792657 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:41:51 crc kubenswrapper[4669]: I1010 09:41:51.310336 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx"] Oct 10 09:41:51 crc kubenswrapper[4669]: I1010 09:41:51.361011 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" event={"ID":"5552238b-e22c-4a4e-b6ed-8ed786a65cbb","Type":"ContainerStarted","Data":"6d5b60456ccf22e974566f36f2c063773195fe6b795039d97e19729f9bdc35dd"} Oct 10 09:41:52 crc kubenswrapper[4669]: I1010 09:41:52.372026 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" event={"ID":"5552238b-e22c-4a4e-b6ed-8ed786a65cbb","Type":"ContainerStarted","Data":"2be0ae918c7b33c3f5920cac2599d93a77329f10a33637cbeb5a98df92fe9853"} Oct 10 09:41:52 crc kubenswrapper[4669]: I1010 09:41:52.400119 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" podStartSLOduration=1.937133147 podStartE2EDuration="2.400092935s" podCreationTimestamp="2025-10-10 09:41:50 +0000 UTC" firstStartedPulling="2025-10-10 09:41:51.318298184 +0000 UTC m=+1854.334316926" lastFinishedPulling="2025-10-10 09:41:51.781257932 +0000 UTC m=+1854.797276714" observedRunningTime="2025-10-10 09:41:52.389694352 +0000 UTC m=+1855.405713134" watchObservedRunningTime="2025-10-10 09:41:52.400092935 +0000 UTC m=+1855.416111707" Oct 10 09:42:01 crc kubenswrapper[4669]: I1010 09:42:01.463551 4669 generic.go:334] "Generic (PLEG): container finished" podID="5552238b-e22c-4a4e-b6ed-8ed786a65cbb" containerID="2be0ae918c7b33c3f5920cac2599d93a77329f10a33637cbeb5a98df92fe9853" exitCode=0 Oct 10 09:42:01 crc kubenswrapper[4669]: I1010 09:42:01.463645 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" event={"ID":"5552238b-e22c-4a4e-b6ed-8ed786a65cbb","Type":"ContainerDied","Data":"2be0ae918c7b33c3f5920cac2599d93a77329f10a33637cbeb5a98df92fe9853"} Oct 10 09:42:02 crc kubenswrapper[4669]: I1010 09:42:02.932640 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:42:02 crc kubenswrapper[4669]: I1010 09:42:02.970620 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spz29\" (UniqueName: \"kubernetes.io/projected/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-kube-api-access-spz29\") pod \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " Oct 10 09:42:02 crc kubenswrapper[4669]: I1010 09:42:02.970787 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-ssh-key\") pod \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " Oct 10 09:42:02 crc kubenswrapper[4669]: I1010 09:42:02.970844 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-inventory\") pod \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\" (UID: \"5552238b-e22c-4a4e-b6ed-8ed786a65cbb\") " Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.002081 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-kube-api-access-spz29" (OuterVolumeSpecName: "kube-api-access-spz29") pod "5552238b-e22c-4a4e-b6ed-8ed786a65cbb" (UID: "5552238b-e22c-4a4e-b6ed-8ed786a65cbb"). InnerVolumeSpecName "kube-api-access-spz29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.032373 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5552238b-e22c-4a4e-b6ed-8ed786a65cbb" (UID: "5552238b-e22c-4a4e-b6ed-8ed786a65cbb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.033296 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-inventory" (OuterVolumeSpecName: "inventory") pod "5552238b-e22c-4a4e-b6ed-8ed786a65cbb" (UID: "5552238b-e22c-4a4e-b6ed-8ed786a65cbb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.073659 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.073727 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.073742 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spz29\" (UniqueName: \"kubernetes.io/projected/5552238b-e22c-4a4e-b6ed-8ed786a65cbb-kube-api-access-spz29\") on node \"crc\" DevicePath \"\"" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.481706 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" event={"ID":"5552238b-e22c-4a4e-b6ed-8ed786a65cbb","Type":"ContainerDied","Data":"6d5b60456ccf22e974566f36f2c063773195fe6b795039d97e19729f9bdc35dd"} Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.481745 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d5b60456ccf22e974566f36f2c063773195fe6b795039d97e19729f9bdc35dd" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.481748 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.550449 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8"] Oct 10 09:42:03 crc kubenswrapper[4669]: E1010 09:42:03.550864 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5552238b-e22c-4a4e-b6ed-8ed786a65cbb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.550884 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5552238b-e22c-4a4e-b6ed-8ed786a65cbb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.551081 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5552238b-e22c-4a4e-b6ed-8ed786a65cbb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.551697 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.554342 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.554701 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.554920 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.555034 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.573751 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8"] Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.582744 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.582799 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.582995 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wbpx\" (UniqueName: \"kubernetes.io/projected/f6566bfc-779b-47b5-941b-88d6fd5d13eb-kube-api-access-4wbpx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.683829 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wbpx\" (UniqueName: \"kubernetes.io/projected/f6566bfc-779b-47b5-941b-88d6fd5d13eb-kube-api-access-4wbpx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.683908 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.683935 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.687781 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.694133 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.707128 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wbpx\" (UniqueName: \"kubernetes.io/projected/f6566bfc-779b-47b5-941b-88d6fd5d13eb-kube-api-access-4wbpx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:03 crc kubenswrapper[4669]: I1010 09:42:03.869230 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:04 crc kubenswrapper[4669]: I1010 09:42:04.561537 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8"] Oct 10 09:42:04 crc kubenswrapper[4669]: W1010 09:42:04.574792 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6566bfc_779b_47b5_941b_88d6fd5d13eb.slice/crio-6381481a2ab2d3157c12d02116a56331b23fc90f49c41edc011a561b7d640a2f WatchSource:0}: Error finding container 6381481a2ab2d3157c12d02116a56331b23fc90f49c41edc011a561b7d640a2f: Status 404 returned error can't find the container with id 6381481a2ab2d3157c12d02116a56331b23fc90f49c41edc011a561b7d640a2f Oct 10 09:42:04 crc kubenswrapper[4669]: I1010 09:42:04.577101 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:42:05 crc kubenswrapper[4669]: I1010 09:42:05.509650 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" event={"ID":"f6566bfc-779b-47b5-941b-88d6fd5d13eb","Type":"ContainerStarted","Data":"c9d89ba1a978092961f06dff1f60abd5e674713a40729484e2ae719d342a1d72"} Oct 10 09:42:05 crc kubenswrapper[4669]: I1010 09:42:05.509978 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" event={"ID":"f6566bfc-779b-47b5-941b-88d6fd5d13eb","Type":"ContainerStarted","Data":"6381481a2ab2d3157c12d02116a56331b23fc90f49c41edc011a561b7d640a2f"} Oct 10 09:42:05 crc kubenswrapper[4669]: I1010 09:42:05.530721 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" podStartSLOduration=2.083385716 podStartE2EDuration="2.530697342s" podCreationTimestamp="2025-10-10 09:42:03 +0000 UTC" firstStartedPulling="2025-10-10 09:42:04.576849852 +0000 UTC m=+1867.592868604" lastFinishedPulling="2025-10-10 09:42:05.024161488 +0000 UTC m=+1868.040180230" observedRunningTime="2025-10-10 09:42:05.521317411 +0000 UTC m=+1868.537336173" watchObservedRunningTime="2025-10-10 09:42:05.530697342 +0000 UTC m=+1868.546716084" Oct 10 09:42:11 crc kubenswrapper[4669]: I1010 09:42:11.395771 4669 scope.go:117] "RemoveContainer" containerID="fa3013e1d7a3d6f01698c1fbabf43c0d5477ab4c837c93f6568e4505fa6dd153" Oct 10 09:42:15 crc kubenswrapper[4669]: I1010 09:42:15.612069 4669 generic.go:334] "Generic (PLEG): container finished" podID="f6566bfc-779b-47b5-941b-88d6fd5d13eb" containerID="c9d89ba1a978092961f06dff1f60abd5e674713a40729484e2ae719d342a1d72" exitCode=0 Oct 10 09:42:15 crc kubenswrapper[4669]: I1010 09:42:15.612215 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" event={"ID":"f6566bfc-779b-47b5-941b-88d6fd5d13eb","Type":"ContainerDied","Data":"c9d89ba1a978092961f06dff1f60abd5e674713a40729484e2ae719d342a1d72"} Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.201817 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.267638 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-ssh-key\") pod \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.267946 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wbpx\" (UniqueName: \"kubernetes.io/projected/f6566bfc-779b-47b5-941b-88d6fd5d13eb-kube-api-access-4wbpx\") pod \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.267980 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-inventory\") pod \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\" (UID: \"f6566bfc-779b-47b5-941b-88d6fd5d13eb\") " Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.272874 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6566bfc-779b-47b5-941b-88d6fd5d13eb-kube-api-access-4wbpx" (OuterVolumeSpecName: "kube-api-access-4wbpx") pod "f6566bfc-779b-47b5-941b-88d6fd5d13eb" (UID: "f6566bfc-779b-47b5-941b-88d6fd5d13eb"). InnerVolumeSpecName "kube-api-access-4wbpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.293919 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6566bfc-779b-47b5-941b-88d6fd5d13eb" (UID: "f6566bfc-779b-47b5-941b-88d6fd5d13eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.304246 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-inventory" (OuterVolumeSpecName: "inventory") pod "f6566bfc-779b-47b5-941b-88d6fd5d13eb" (UID: "f6566bfc-779b-47b5-941b-88d6fd5d13eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.369377 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.369410 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wbpx\" (UniqueName: \"kubernetes.io/projected/f6566bfc-779b-47b5-941b-88d6fd5d13eb-kube-api-access-4wbpx\") on node \"crc\" DevicePath \"\"" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.369422 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6566bfc-779b-47b5-941b-88d6fd5d13eb-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.628174 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" event={"ID":"f6566bfc-779b-47b5-941b-88d6fd5d13eb","Type":"ContainerDied","Data":"6381481a2ab2d3157c12d02116a56331b23fc90f49c41edc011a561b7d640a2f"} Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.628210 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6381481a2ab2d3157c12d02116a56331b23fc90f49c41edc011a561b7d640a2f" Oct 10 09:42:17 crc kubenswrapper[4669]: I1010 09:42:17.628455 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8" Oct 10 09:43:54 crc kubenswrapper[4669]: I1010 09:43:54.275461 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:43:54 crc kubenswrapper[4669]: I1010 09:43:54.276208 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:44:24 crc kubenswrapper[4669]: I1010 09:44:24.275328 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:44:24 crc kubenswrapper[4669]: I1010 09:44:24.276143 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:44:54 crc kubenswrapper[4669]: I1010 09:44:54.274489 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:44:54 crc kubenswrapper[4669]: I1010 09:44:54.275228 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:44:54 crc kubenswrapper[4669]: I1010 09:44:54.275279 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:44:54 crc kubenswrapper[4669]: I1010 09:44:54.276181 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c8ef3ad011601a42242c3a0f4bbb5ca2056df6593c3e51f7cd342be4e813559"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:44:54 crc kubenswrapper[4669]: I1010 09:44:54.276275 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://3c8ef3ad011601a42242c3a0f4bbb5ca2056df6593c3e51f7cd342be4e813559" gracePeriod=600 Oct 10 09:44:55 crc kubenswrapper[4669]: I1010 09:44:55.123351 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="3c8ef3ad011601a42242c3a0f4bbb5ca2056df6593c3e51f7cd342be4e813559" exitCode=0 Oct 10 09:44:55 crc kubenswrapper[4669]: I1010 09:44:55.123637 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"3c8ef3ad011601a42242c3a0f4bbb5ca2056df6593c3e51f7cd342be4e813559"} Oct 10 09:44:55 crc kubenswrapper[4669]: I1010 09:44:55.123966 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf"} Oct 10 09:44:55 crc kubenswrapper[4669]: I1010 09:44:55.123994 4669 scope.go:117] "RemoveContainer" containerID="80b33f972cc081073f3cae2287688626b94c852829047871754b61e80a920fd9" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.219502 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qx48p"] Oct 10 09:44:58 crc kubenswrapper[4669]: E1010 09:44:58.220781 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6566bfc-779b-47b5-941b-88d6fd5d13eb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.220810 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6566bfc-779b-47b5-941b-88d6fd5d13eb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.221102 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6566bfc-779b-47b5-941b-88d6fd5d13eb" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.223165 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.229188 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qx48p"] Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.379099 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnpdc\" (UniqueName: \"kubernetes.io/projected/0906a875-76ed-4f6b-91f0-035c41926dcd-kube-api-access-rnpdc\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.379386 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-utilities\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.379450 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-catalog-content\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.481198 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnpdc\" (UniqueName: \"kubernetes.io/projected/0906a875-76ed-4f6b-91f0-035c41926dcd-kube-api-access-rnpdc\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.481281 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-utilities\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.481323 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-catalog-content\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.481796 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-catalog-content\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.482315 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-utilities\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.502562 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnpdc\" (UniqueName: \"kubernetes.io/projected/0906a875-76ed-4f6b-91f0-035c41926dcd-kube-api-access-rnpdc\") pod \"community-operators-qx48p\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:58 crc kubenswrapper[4669]: I1010 09:44:58.558021 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:44:59 crc kubenswrapper[4669]: I1010 09:44:59.089884 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qx48p"] Oct 10 09:44:59 crc kubenswrapper[4669]: I1010 09:44:59.160991 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qx48p" event={"ID":"0906a875-76ed-4f6b-91f0-035c41926dcd","Type":"ContainerStarted","Data":"84fee2da20a1abc0f12103c9f5048bfce8bdf42e5aee4fc696a494f7c2cd1232"} Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.155396 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46"] Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.161592 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.167983 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.170712 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.179966 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46"] Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.190541 4669 generic.go:334] "Generic (PLEG): container finished" podID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerID="f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65" exitCode=0 Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.190606 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qx48p" event={"ID":"0906a875-76ed-4f6b-91f0-035c41926dcd","Type":"ContainerDied","Data":"f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65"} Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.216901 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-config-volume\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.216977 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppjs9\" (UniqueName: \"kubernetes.io/projected/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-kube-api-access-ppjs9\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.217025 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-secret-volume\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.319208 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-config-volume\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.319531 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppjs9\" (UniqueName: \"kubernetes.io/projected/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-kube-api-access-ppjs9\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.319555 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-secret-volume\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.320726 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-config-volume\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.331971 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-secret-volume\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.336348 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppjs9\" (UniqueName: \"kubernetes.io/projected/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-kube-api-access-ppjs9\") pod \"collect-profiles-29334825-7sv46\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.501984 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:00 crc kubenswrapper[4669]: I1010 09:45:00.960922 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46"] Oct 10 09:45:01 crc kubenswrapper[4669]: I1010 09:45:01.207040 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qx48p" event={"ID":"0906a875-76ed-4f6b-91f0-035c41926dcd","Type":"ContainerStarted","Data":"aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36"} Oct 10 09:45:01 crc kubenswrapper[4669]: I1010 09:45:01.208723 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" event={"ID":"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0","Type":"ContainerStarted","Data":"75503a8cddbf3b864ad271e340a446e3e4ae36027424a41ae29a4513158b7116"} Oct 10 09:45:01 crc kubenswrapper[4669]: I1010 09:45:01.208784 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" event={"ID":"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0","Type":"ContainerStarted","Data":"ab7c6217a1d56c16e17e269c199a0dcfcd3828c0b80503af99aa4df440b8606a"} Oct 10 09:45:01 crc kubenswrapper[4669]: I1010 09:45:01.258376 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" podStartSLOduration=1.258355523 podStartE2EDuration="1.258355523s" podCreationTimestamp="2025-10-10 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 09:45:01.252361162 +0000 UTC m=+2044.268379904" watchObservedRunningTime="2025-10-10 09:45:01.258355523 +0000 UTC m=+2044.274374265" Oct 10 09:45:02 crc kubenswrapper[4669]: I1010 09:45:02.223327 4669 generic.go:334] "Generic (PLEG): container finished" podID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerID="aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36" exitCode=0 Oct 10 09:45:02 crc kubenswrapper[4669]: I1010 09:45:02.223467 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qx48p" event={"ID":"0906a875-76ed-4f6b-91f0-035c41926dcd","Type":"ContainerDied","Data":"aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36"} Oct 10 09:45:02 crc kubenswrapper[4669]: I1010 09:45:02.230279 4669 generic.go:334] "Generic (PLEG): container finished" podID="1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" containerID="75503a8cddbf3b864ad271e340a446e3e4ae36027424a41ae29a4513158b7116" exitCode=0 Oct 10 09:45:02 crc kubenswrapper[4669]: I1010 09:45:02.230320 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" event={"ID":"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0","Type":"ContainerDied","Data":"75503a8cddbf3b864ad271e340a446e3e4ae36027424a41ae29a4513158b7116"} Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.240615 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qx48p" event={"ID":"0906a875-76ed-4f6b-91f0-035c41926dcd","Type":"ContainerStarted","Data":"88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696"} Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.267218 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qx48p" podStartSLOduration=2.61771057 podStartE2EDuration="5.267190695s" podCreationTimestamp="2025-10-10 09:44:58 +0000 UTC" firstStartedPulling="2025-10-10 09:45:00.192869195 +0000 UTC m=+2043.208887937" lastFinishedPulling="2025-10-10 09:45:02.84234931 +0000 UTC m=+2045.858368062" observedRunningTime="2025-10-10 09:45:03.258749065 +0000 UTC m=+2046.274767817" watchObservedRunningTime="2025-10-10 09:45:03.267190695 +0000 UTC m=+2046.283209457" Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.561527 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.679559 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-config-volume\") pod \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.679637 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-secret-volume\") pod \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.679849 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppjs9\" (UniqueName: \"kubernetes.io/projected/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-kube-api-access-ppjs9\") pod \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\" (UID: \"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0\") " Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.680780 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" (UID: "1442cabb-eb6c-4a8a-9c00-d272bdecc2f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.685244 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-kube-api-access-ppjs9" (OuterVolumeSpecName: "kube-api-access-ppjs9") pod "1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" (UID: "1442cabb-eb6c-4a8a-9c00-d272bdecc2f0"). InnerVolumeSpecName "kube-api-access-ppjs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.697244 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" (UID: "1442cabb-eb6c-4a8a-9c00-d272bdecc2f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.782791 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppjs9\" (UniqueName: \"kubernetes.io/projected/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-kube-api-access-ppjs9\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.782857 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:03 crc kubenswrapper[4669]: I1010 09:45:03.782872 4669 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:04 crc kubenswrapper[4669]: I1010 09:45:04.254977 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" event={"ID":"1442cabb-eb6c-4a8a-9c00-d272bdecc2f0","Type":"ContainerDied","Data":"ab7c6217a1d56c16e17e269c199a0dcfcd3828c0b80503af99aa4df440b8606a"} Oct 10 09:45:04 crc kubenswrapper[4669]: I1010 09:45:04.255027 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab7c6217a1d56c16e17e269c199a0dcfcd3828c0b80503af99aa4df440b8606a" Oct 10 09:45:04 crc kubenswrapper[4669]: I1010 09:45:04.254990 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46" Oct 10 09:45:04 crc kubenswrapper[4669]: I1010 09:45:04.335742 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79"] Oct 10 09:45:04 crc kubenswrapper[4669]: I1010 09:45:04.342843 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334780-xjz79"] Oct 10 09:45:05 crc kubenswrapper[4669]: I1010 09:45:05.810569 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98efe68d-5663-488a-92ca-c02a288b524f" path="/var/lib/kubelet/pods/98efe68d-5663-488a-92ca-c02a288b524f/volumes" Oct 10 09:45:08 crc kubenswrapper[4669]: I1010 09:45:08.559609 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:45:08 crc kubenswrapper[4669]: I1010 09:45:08.561009 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:45:08 crc kubenswrapper[4669]: I1010 09:45:08.631856 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:45:09 crc kubenswrapper[4669]: I1010 09:45:09.353083 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:45:09 crc kubenswrapper[4669]: I1010 09:45:09.407506 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qx48p"] Oct 10 09:45:11 crc kubenswrapper[4669]: I1010 09:45:11.306051 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qx48p" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="registry-server" containerID="cri-o://88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696" gracePeriod=2 Oct 10 09:45:11 crc kubenswrapper[4669]: E1010 09:45:11.504728 4669 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0906a875_76ed_4f6b_91f0_035c41926dcd.slice/crio-conmon-88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696.scope\": RecentStats: unable to find data in memory cache]" Oct 10 09:45:11 crc kubenswrapper[4669]: I1010 09:45:11.556271 4669 scope.go:117] "RemoveContainer" containerID="f8edf14c55d21fe2700b9025389717cb6da78032ec4ae2098d6d03ed8542851b" Oct 10 09:45:11 crc kubenswrapper[4669]: I1010 09:45:11.859018 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.049823 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-catalog-content\") pod \"0906a875-76ed-4f6b-91f0-035c41926dcd\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.049937 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-utilities\") pod \"0906a875-76ed-4f6b-91f0-035c41926dcd\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.050064 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnpdc\" (UniqueName: \"kubernetes.io/projected/0906a875-76ed-4f6b-91f0-035c41926dcd-kube-api-access-rnpdc\") pod \"0906a875-76ed-4f6b-91f0-035c41926dcd\" (UID: \"0906a875-76ed-4f6b-91f0-035c41926dcd\") " Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.050631 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-utilities" (OuterVolumeSpecName: "utilities") pod "0906a875-76ed-4f6b-91f0-035c41926dcd" (UID: "0906a875-76ed-4f6b-91f0-035c41926dcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.062815 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0906a875-76ed-4f6b-91f0-035c41926dcd-kube-api-access-rnpdc" (OuterVolumeSpecName: "kube-api-access-rnpdc") pod "0906a875-76ed-4f6b-91f0-035c41926dcd" (UID: "0906a875-76ed-4f6b-91f0-035c41926dcd"). InnerVolumeSpecName "kube-api-access-rnpdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.098151 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0906a875-76ed-4f6b-91f0-035c41926dcd" (UID: "0906a875-76ed-4f6b-91f0-035c41926dcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.151984 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.152017 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnpdc\" (UniqueName: \"kubernetes.io/projected/0906a875-76ed-4f6b-91f0-035c41926dcd-kube-api-access-rnpdc\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.152029 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0906a875-76ed-4f6b-91f0-035c41926dcd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.319655 4669 generic.go:334] "Generic (PLEG): container finished" podID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerID="88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696" exitCode=0 Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.319705 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qx48p" event={"ID":"0906a875-76ed-4f6b-91f0-035c41926dcd","Type":"ContainerDied","Data":"88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696"} Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.319758 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qx48p" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.319773 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qx48p" event={"ID":"0906a875-76ed-4f6b-91f0-035c41926dcd","Type":"ContainerDied","Data":"84fee2da20a1abc0f12103c9f5048bfce8bdf42e5aee4fc696a494f7c2cd1232"} Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.319821 4669 scope.go:117] "RemoveContainer" containerID="88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.341613 4669 scope.go:117] "RemoveContainer" containerID="aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.370594 4669 scope.go:117] "RemoveContainer" containerID="f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.378837 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qx48p"] Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.385190 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qx48p"] Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.389640 4669 scope.go:117] "RemoveContainer" containerID="88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696" Oct 10 09:45:12 crc kubenswrapper[4669]: E1010 09:45:12.390067 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696\": container with ID starting with 88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696 not found: ID does not exist" containerID="88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.390097 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696"} err="failed to get container status \"88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696\": rpc error: code = NotFound desc = could not find container \"88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696\": container with ID starting with 88f278f4398aee41efe8a5131fca5eb48953dbc24467efa96c58c79febf3f696 not found: ID does not exist" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.390119 4669 scope.go:117] "RemoveContainer" containerID="aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36" Oct 10 09:45:12 crc kubenswrapper[4669]: E1010 09:45:12.390365 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36\": container with ID starting with aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36 not found: ID does not exist" containerID="aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.390389 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36"} err="failed to get container status \"aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36\": rpc error: code = NotFound desc = could not find container \"aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36\": container with ID starting with aab6bebc3dad81cbcac7745ac7f633e15362f53a6c9c04272dc9f48191965e36 not found: ID does not exist" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.390408 4669 scope.go:117] "RemoveContainer" containerID="f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65" Oct 10 09:45:12 crc kubenswrapper[4669]: E1010 09:45:12.390658 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65\": container with ID starting with f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65 not found: ID does not exist" containerID="f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65" Oct 10 09:45:12 crc kubenswrapper[4669]: I1010 09:45:12.390680 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65"} err="failed to get container status \"f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65\": rpc error: code = NotFound desc = could not find container \"f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65\": container with ID starting with f480b98a4a93549e6f6b5c11aa61447764b4413bb81e8cab4611e179f7992b65 not found: ID does not exist" Oct 10 09:45:13 crc kubenswrapper[4669]: I1010 09:45:13.815135 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" path="/var/lib/kubelet/pods/0906a875-76ed-4f6b-91f0-035c41926dcd/volumes" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.982744 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mr6zn"] Oct 10 09:45:44 crc kubenswrapper[4669]: E1010 09:45:44.984405 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="extract-content" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.984427 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="extract-content" Oct 10 09:45:44 crc kubenswrapper[4669]: E1010 09:45:44.984488 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="registry-server" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.984640 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="registry-server" Oct 10 09:45:44 crc kubenswrapper[4669]: E1010 09:45:44.984662 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="extract-utilities" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.984674 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="extract-utilities" Oct 10 09:45:44 crc kubenswrapper[4669]: E1010 09:45:44.984695 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" containerName="collect-profiles" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.984707 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" containerName="collect-profiles" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.985245 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" containerName="collect-profiles" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.985278 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0906a875-76ed-4f6b-91f0-035c41926dcd" containerName="registry-server" Oct 10 09:45:44 crc kubenswrapper[4669]: I1010 09:45:44.992938 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.007887 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mr6zn"] Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.172874 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-utilities\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.172987 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-catalog-content\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.173097 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z49b\" (UniqueName: \"kubernetes.io/projected/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-kube-api-access-4z49b\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.274445 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-utilities\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.274528 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-catalog-content\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.274575 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z49b\" (UniqueName: \"kubernetes.io/projected/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-kube-api-access-4z49b\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.275048 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-utilities\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.275298 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-catalog-content\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.306699 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z49b\" (UniqueName: \"kubernetes.io/projected/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-kube-api-access-4z49b\") pod \"redhat-marketplace-mr6zn\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.327534 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:45 crc kubenswrapper[4669]: I1010 09:45:45.810062 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mr6zn"] Oct 10 09:45:46 crc kubenswrapper[4669]: I1010 09:45:46.610991 4669 generic.go:334] "Generic (PLEG): container finished" podID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerID="6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7" exitCode=0 Oct 10 09:45:46 crc kubenswrapper[4669]: I1010 09:45:46.611281 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mr6zn" event={"ID":"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9","Type":"ContainerDied","Data":"6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7"} Oct 10 09:45:46 crc kubenswrapper[4669]: I1010 09:45:46.611316 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mr6zn" event={"ID":"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9","Type":"ContainerStarted","Data":"f1ee027ac93f3470c756d8cf78739df5d979a628e3a2c7e5dc1e2f74392ac4e0"} Oct 10 09:45:47 crc kubenswrapper[4669]: I1010 09:45:47.621058 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mr6zn" event={"ID":"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9","Type":"ContainerStarted","Data":"05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072"} Oct 10 09:45:48 crc kubenswrapper[4669]: I1010 09:45:48.630572 4669 generic.go:334] "Generic (PLEG): container finished" podID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerID="05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072" exitCode=0 Oct 10 09:45:48 crc kubenswrapper[4669]: I1010 09:45:48.630679 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mr6zn" event={"ID":"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9","Type":"ContainerDied","Data":"05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072"} Oct 10 09:45:49 crc kubenswrapper[4669]: I1010 09:45:49.644802 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mr6zn" event={"ID":"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9","Type":"ContainerStarted","Data":"ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da"} Oct 10 09:45:49 crc kubenswrapper[4669]: I1010 09:45:49.675349 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mr6zn" podStartSLOduration=3.246870073 podStartE2EDuration="5.675313294s" podCreationTimestamp="2025-10-10 09:45:44 +0000 UTC" firstStartedPulling="2025-10-10 09:45:46.61483514 +0000 UTC m=+2089.630853882" lastFinishedPulling="2025-10-10 09:45:49.043278361 +0000 UTC m=+2092.059297103" observedRunningTime="2025-10-10 09:45:49.667447152 +0000 UTC m=+2092.683465904" watchObservedRunningTime="2025-10-10 09:45:49.675313294 +0000 UTC m=+2092.691332046" Oct 10 09:45:55 crc kubenswrapper[4669]: I1010 09:45:55.328639 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:55 crc kubenswrapper[4669]: I1010 09:45:55.329273 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:55 crc kubenswrapper[4669]: I1010 09:45:55.376076 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:55 crc kubenswrapper[4669]: I1010 09:45:55.746999 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:55 crc kubenswrapper[4669]: I1010 09:45:55.808156 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mr6zn"] Oct 10 09:45:57 crc kubenswrapper[4669]: I1010 09:45:57.712395 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mr6zn" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="registry-server" containerID="cri-o://ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da" gracePeriod=2 Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.157366 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.224875 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-catalog-content\") pod \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.224934 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-utilities\") pod \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.225000 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z49b\" (UniqueName: \"kubernetes.io/projected/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-kube-api-access-4z49b\") pod \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\" (UID: \"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9\") " Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.225819 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-utilities" (OuterVolumeSpecName: "utilities") pod "0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" (UID: "0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.233729 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-kube-api-access-4z49b" (OuterVolumeSpecName: "kube-api-access-4z49b") pod "0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" (UID: "0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9"). InnerVolumeSpecName "kube-api-access-4z49b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.241505 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" (UID: "0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.326476 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.326512 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.326522 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z49b\" (UniqueName: \"kubernetes.io/projected/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9-kube-api-access-4z49b\") on node \"crc\" DevicePath \"\"" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.723106 4669 generic.go:334] "Generic (PLEG): container finished" podID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerID="ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da" exitCode=0 Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.723153 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mr6zn" event={"ID":"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9","Type":"ContainerDied","Data":"ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da"} Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.723182 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mr6zn" event={"ID":"0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9","Type":"ContainerDied","Data":"f1ee027ac93f3470c756d8cf78739df5d979a628e3a2c7e5dc1e2f74392ac4e0"} Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.723182 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mr6zn" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.723202 4669 scope.go:117] "RemoveContainer" containerID="ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.744955 4669 scope.go:117] "RemoveContainer" containerID="05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.769483 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mr6zn"] Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.778739 4669 scope.go:117] "RemoveContainer" containerID="6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.784698 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mr6zn"] Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.818696 4669 scope.go:117] "RemoveContainer" containerID="ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da" Oct 10 09:45:58 crc kubenswrapper[4669]: E1010 09:45:58.819268 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da\": container with ID starting with ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da not found: ID does not exist" containerID="ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.819303 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da"} err="failed to get container status \"ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da\": rpc error: code = NotFound desc = could not find container \"ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da\": container with ID starting with ce8c41c9f14a057a9df5d6e99dbb285a371143ee97e4c994e87df94fc67f34da not found: ID does not exist" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.819346 4669 scope.go:117] "RemoveContainer" containerID="05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072" Oct 10 09:45:58 crc kubenswrapper[4669]: E1010 09:45:58.819873 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072\": container with ID starting with 05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072 not found: ID does not exist" containerID="05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.819920 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072"} err="failed to get container status \"05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072\": rpc error: code = NotFound desc = could not find container \"05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072\": container with ID starting with 05d51ad03df0e75f481b0c6cd14c688a38427efeec067a0c42f9a026475cf072 not found: ID does not exist" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.819937 4669 scope.go:117] "RemoveContainer" containerID="6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7" Oct 10 09:45:58 crc kubenswrapper[4669]: E1010 09:45:58.820370 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7\": container with ID starting with 6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7 not found: ID does not exist" containerID="6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7" Oct 10 09:45:58 crc kubenswrapper[4669]: I1010 09:45:58.820417 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7"} err="failed to get container status \"6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7\": rpc error: code = NotFound desc = could not find container \"6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7\": container with ID starting with 6cf8d14436814ee2f14ab454258ebd1213fab5a880b632ee4cd19affb9ededa7 not found: ID does not exist" Oct 10 09:45:59 crc kubenswrapper[4669]: I1010 09:45:59.806525 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" path="/var/lib/kubelet/pods/0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9/volumes" Oct 10 09:46:54 crc kubenswrapper[4669]: I1010 09:46:54.274622 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:46:54 crc kubenswrapper[4669]: I1010 09:46:54.275288 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:47:24 crc kubenswrapper[4669]: I1010 09:47:24.275704 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:47:24 crc kubenswrapper[4669]: I1010 09:47:24.276361 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.836288 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.841839 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.847054 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.853665 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mmkmx"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.861267 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-n2bvp"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.872262 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-wq2fj"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.881208 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb99"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.887180 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.892260 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.898431 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.907052 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.914956 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.920984 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.927924 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cmb99"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.935007 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-6mhmh"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.943180 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pd9fp"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.949898 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.956087 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vz7sz"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.962703 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-g5w67"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.968768 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-v5cl8"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.978612 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxp2l"] Oct 10 09:47:50 crc kubenswrapper[4669]: I1010 09:47:50.984885 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hw5nv"] Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.850956 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1934fe86-0d23-4abb-be37-03beb45bd95d" path="/var/lib/kubelet/pods/1934fe86-0d23-4abb-be37-03beb45bd95d/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.851481 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4505ce35-c314-4aef-b26f-821513df946f" path="/var/lib/kubelet/pods/4505ce35-c314-4aef-b26f-821513df946f/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.852014 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5552238b-e22c-4a4e-b6ed-8ed786a65cbb" path="/var/lib/kubelet/pods/5552238b-e22c-4a4e-b6ed-8ed786a65cbb/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.852490 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94" path="/var/lib/kubelet/pods/87088e0f-7d9f-41aa-9eeb-c3d82d4ceb94/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.853444 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c21d000-f125-4187-bafc-75c372f943a1" path="/var/lib/kubelet/pods/8c21d000-f125-4187-bafc-75c372f943a1/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.853961 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b814d616-5da3-4efb-844f-854a6ce7d878" path="/var/lib/kubelet/pods/b814d616-5da3-4efb-844f-854a6ce7d878/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.854541 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf129287-698a-41fd-8623-83628590e428" path="/var/lib/kubelet/pods/bf129287-698a-41fd-8623-83628590e428/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.855759 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57" path="/var/lib/kubelet/pods/cf5f2fc9-b211-4f0f-a0e9-d78ce32aaf57/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.856267 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e253a578-2a65-406b-8b9d-3128c12c2cdb" path="/var/lib/kubelet/pods/e253a578-2a65-406b-8b9d-3128c12c2cdb/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.856767 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4f5ee49-2ac8-49cf-b570-eaa3ad05c148" path="/var/lib/kubelet/pods/f4f5ee49-2ac8-49cf-b570-eaa3ad05c148/volumes" Oct 10 09:47:51 crc kubenswrapper[4669]: I1010 09:47:51.858027 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6566bfc-779b-47b5-941b-88d6fd5d13eb" path="/var/lib/kubelet/pods/f6566bfc-779b-47b5-941b-88d6fd5d13eb/volumes" Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.275278 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.275553 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.275611 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.276289 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.276334 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" gracePeriod=600 Oct 10 09:47:54 crc kubenswrapper[4669]: E1010 09:47:54.397216 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.778279 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" exitCode=0 Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.778340 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf"} Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.778401 4669 scope.go:117] "RemoveContainer" containerID="3c8ef3ad011601a42242c3a0f4bbb5ca2056df6593c3e51f7cd342be4e813559" Oct 10 09:47:54 crc kubenswrapper[4669]: I1010 09:47:54.779687 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:47:54 crc kubenswrapper[4669]: E1010 09:47:54.780444 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.948915 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh"] Oct 10 09:48:04 crc kubenswrapper[4669]: E1010 09:48:04.949737 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="registry-server" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.949753 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="registry-server" Oct 10 09:48:04 crc kubenswrapper[4669]: E1010 09:48:04.949768 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="extract-content" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.949776 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="extract-content" Oct 10 09:48:04 crc kubenswrapper[4669]: E1010 09:48:04.949785 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="extract-utilities" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.949791 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="extract-utilities" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.949961 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aac9ea4-16d6-4f06-aa1f-5d1e73ab4cb9" containerName="registry-server" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.950651 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.954086 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.954336 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.954456 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.958335 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.961713 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:48:04 crc kubenswrapper[4669]: I1010 09:48:04.961832 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh"] Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.037439 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.037649 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.037699 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.037829 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq9km\" (UniqueName: \"kubernetes.io/projected/5c0a91b3-706e-47df-8f2b-84733ad00593-kube-api-access-qq9km\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.037945 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.139938 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.140321 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.140875 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq9km\" (UniqueName: \"kubernetes.io/projected/5c0a91b3-706e-47df-8f2b-84733ad00593-kube-api-access-qq9km\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.140946 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.140999 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.145676 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.148202 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.150053 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.158404 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.159218 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq9km\" (UniqueName: \"kubernetes.io/projected/5c0a91b3-706e-47df-8f2b-84733ad00593-kube-api-access-qq9km\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.272003 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.852702 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh"] Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.869849 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:48:05 crc kubenswrapper[4669]: I1010 09:48:05.893159 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" event={"ID":"5c0a91b3-706e-47df-8f2b-84733ad00593","Type":"ContainerStarted","Data":"81a39cf696b38a3d65a45d9456b29e40ccc509ef0ef5eb6e91b06154458b2fb7"} Oct 10 09:48:06 crc kubenswrapper[4669]: I1010 09:48:06.795940 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:48:06 crc kubenswrapper[4669]: E1010 09:48:06.796563 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:48:06 crc kubenswrapper[4669]: I1010 09:48:06.904962 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" event={"ID":"5c0a91b3-706e-47df-8f2b-84733ad00593","Type":"ContainerStarted","Data":"f609c614e3938354b1bd2dff0c441cb9f074f20df2bfe4851d12eefe73462e6e"} Oct 10 09:48:11 crc kubenswrapper[4669]: I1010 09:48:11.772719 4669 scope.go:117] "RemoveContainer" containerID="2be0ae918c7b33c3f5920cac2599d93a77329f10a33637cbeb5a98df92fe9853" Oct 10 09:48:11 crc kubenswrapper[4669]: I1010 09:48:11.840205 4669 scope.go:117] "RemoveContainer" containerID="e08600586c8017a0e8c5f7d87bfd6f1c8641110f86a20271453e5878e6a3cb35" Oct 10 09:48:11 crc kubenswrapper[4669]: I1010 09:48:11.911834 4669 scope.go:117] "RemoveContainer" containerID="66e1ed52402bd6ce43e45c69541ce9f7a9bdd8fefb2362e54dfc2a36a0ab1f02" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.076571 4669 scope.go:117] "RemoveContainer" containerID="01ada31687f73fea7b0ccec2719af3a7f1b8022d7926549b3a45cfa0538a3253" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.143082 4669 scope.go:117] "RemoveContainer" containerID="0df982b51daa067ddc60b0449155ca31fc67fafc8f809a0cac7d28c14734cbda" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.201547 4669 scope.go:117] "RemoveContainer" containerID="25d9022310ee71ba3703d87ff416aa4d5a3ff840a1e9f29959aa518c4ddde469" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.252335 4669 scope.go:117] "RemoveContainer" containerID="a11a8bd3394ecf53974839036494f0350151dcfca37e4dbc7cab550b63b8df1b" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.286312 4669 scope.go:117] "RemoveContainer" containerID="52c405f6cd5452db64620bb6f5cc4ec390feb43bb43501685970752ac0f0bb82" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.320511 4669 scope.go:117] "RemoveContainer" containerID="a9e6be0de9ee8269114d300be5825c5e168fd33df286bfb9dbe3d6b3dd594302" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.375980 4669 scope.go:117] "RemoveContainer" containerID="c9d89ba1a978092961f06dff1f60abd5e674713a40729484e2ae719d342a1d72" Oct 10 09:48:12 crc kubenswrapper[4669]: I1010 09:48:12.413863 4669 scope.go:117] "RemoveContainer" containerID="66cb8d5f3b05585c190d8238874135c2ba38b9d5538df9755820b58eaed6165b" Oct 10 09:48:19 crc kubenswrapper[4669]: I1010 09:48:19.010238 4669 generic.go:334] "Generic (PLEG): container finished" podID="5c0a91b3-706e-47df-8f2b-84733ad00593" containerID="f609c614e3938354b1bd2dff0c441cb9f074f20df2bfe4851d12eefe73462e6e" exitCode=0 Oct 10 09:48:19 crc kubenswrapper[4669]: I1010 09:48:19.010312 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" event={"ID":"5c0a91b3-706e-47df-8f2b-84733ad00593","Type":"ContainerDied","Data":"f609c614e3938354b1bd2dff0c441cb9f074f20df2bfe4851d12eefe73462e6e"} Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.419236 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.532634 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ssh-key\") pod \"5c0a91b3-706e-47df-8f2b-84733ad00593\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.532799 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-repo-setup-combined-ca-bundle\") pod \"5c0a91b3-706e-47df-8f2b-84733ad00593\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.533557 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-inventory\") pod \"5c0a91b3-706e-47df-8f2b-84733ad00593\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.533770 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ceph\") pod \"5c0a91b3-706e-47df-8f2b-84733ad00593\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.533846 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq9km\" (UniqueName: \"kubernetes.io/projected/5c0a91b3-706e-47df-8f2b-84733ad00593-kube-api-access-qq9km\") pod \"5c0a91b3-706e-47df-8f2b-84733ad00593\" (UID: \"5c0a91b3-706e-47df-8f2b-84733ad00593\") " Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.556825 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ceph" (OuterVolumeSpecName: "ceph") pod "5c0a91b3-706e-47df-8f2b-84733ad00593" (UID: "5c0a91b3-706e-47df-8f2b-84733ad00593"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.556942 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5c0a91b3-706e-47df-8f2b-84733ad00593" (UID: "5c0a91b3-706e-47df-8f2b-84733ad00593"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.557073 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0a91b3-706e-47df-8f2b-84733ad00593-kube-api-access-qq9km" (OuterVolumeSpecName: "kube-api-access-qq9km") pod "5c0a91b3-706e-47df-8f2b-84733ad00593" (UID: "5c0a91b3-706e-47df-8f2b-84733ad00593"). InnerVolumeSpecName "kube-api-access-qq9km". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.572737 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5c0a91b3-706e-47df-8f2b-84733ad00593" (UID: "5c0a91b3-706e-47df-8f2b-84733ad00593"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.636461 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.636697 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq9km\" (UniqueName: \"kubernetes.io/projected/5c0a91b3-706e-47df-8f2b-84733ad00593-kube-api-access-qq9km\") on node \"crc\" DevicePath \"\"" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.636713 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.636725 4669 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.667322 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-inventory" (OuterVolumeSpecName: "inventory") pod "5c0a91b3-706e-47df-8f2b-84733ad00593" (UID: "5c0a91b3-706e-47df-8f2b-84733ad00593"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.738885 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5c0a91b3-706e-47df-8f2b-84733ad00593-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:48:20 crc kubenswrapper[4669]: I1010 09:48:20.795518 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:48:20 crc kubenswrapper[4669]: E1010 09:48:20.795912 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.028025 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" event={"ID":"5c0a91b3-706e-47df-8f2b-84733ad00593","Type":"ContainerDied","Data":"81a39cf696b38a3d65a45d9456b29e40ccc509ef0ef5eb6e91b06154458b2fb7"} Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.028063 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81a39cf696b38a3d65a45d9456b29e40ccc509ef0ef5eb6e91b06154458b2fb7" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.028062 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.119556 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v"] Oct 10 09:48:21 crc kubenswrapper[4669]: E1010 09:48:21.120003 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0a91b3-706e-47df-8f2b-84733ad00593" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.120022 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0a91b3-706e-47df-8f2b-84733ad00593" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.120242 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0a91b3-706e-47df-8f2b-84733ad00593" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.120935 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.123090 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.123300 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.123526 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.123722 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.126517 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.135803 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v"] Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.247926 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6zqm\" (UniqueName: \"kubernetes.io/projected/66dbecf2-f169-4c6c-b8be-efbab612ace8-kube-api-access-z6zqm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.247998 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.248180 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.248357 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.248393 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.350439 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.350539 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.350569 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.350671 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6zqm\" (UniqueName: \"kubernetes.io/projected/66dbecf2-f169-4c6c-b8be-efbab612ace8-kube-api-access-z6zqm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.350732 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.355966 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.358987 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.362038 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.370177 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.373780 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6zqm\" (UniqueName: \"kubernetes.io/projected/66dbecf2-f169-4c6c-b8be-efbab612ace8-kube-api-access-z6zqm\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:21 crc kubenswrapper[4669]: I1010 09:48:21.436417 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:48:22 crc kubenswrapper[4669]: I1010 09:48:22.022731 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v"] Oct 10 09:48:22 crc kubenswrapper[4669]: I1010 09:48:22.038461 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" event={"ID":"66dbecf2-f169-4c6c-b8be-efbab612ace8","Type":"ContainerStarted","Data":"c12f6791ac6499a07e679fb7e2a4556ba21d21c166a671627a8fc8a7fbd03bdb"} Oct 10 09:48:23 crc kubenswrapper[4669]: I1010 09:48:23.047149 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" event={"ID":"66dbecf2-f169-4c6c-b8be-efbab612ace8","Type":"ContainerStarted","Data":"b4a689b2ef12140a336b76e11541a7e15ba6adb45cd65510a30ccfe5da9648b9"} Oct 10 09:48:23 crc kubenswrapper[4669]: I1010 09:48:23.062408 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" podStartSLOduration=1.562046273 podStartE2EDuration="2.062387108s" podCreationTimestamp="2025-10-10 09:48:21 +0000 UTC" firstStartedPulling="2025-10-10 09:48:22.031397086 +0000 UTC m=+2245.047415848" lastFinishedPulling="2025-10-10 09:48:22.531737941 +0000 UTC m=+2245.547756683" observedRunningTime="2025-10-10 09:48:23.061286023 +0000 UTC m=+2246.077304775" watchObservedRunningTime="2025-10-10 09:48:23.062387108 +0000 UTC m=+2246.078405850" Oct 10 09:48:31 crc kubenswrapper[4669]: I1010 09:48:31.797309 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:48:31 crc kubenswrapper[4669]: E1010 09:48:31.798864 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:48:45 crc kubenswrapper[4669]: I1010 09:48:45.796033 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:48:45 crc kubenswrapper[4669]: E1010 09:48:45.796904 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:48:58 crc kubenswrapper[4669]: I1010 09:48:58.795905 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:48:58 crc kubenswrapper[4669]: E1010 09:48:58.797111 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:49:12 crc kubenswrapper[4669]: I1010 09:49:12.795976 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:49:12 crc kubenswrapper[4669]: E1010 09:49:12.796833 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:49:24 crc kubenswrapper[4669]: I1010 09:49:24.795726 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:49:24 crc kubenswrapper[4669]: E1010 09:49:24.796750 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:49:36 crc kubenswrapper[4669]: I1010 09:49:36.796187 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:49:36 crc kubenswrapper[4669]: E1010 09:49:36.797059 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.465760 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vzkls"] Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.468322 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.479404 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzkls"] Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.539684 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-catalog-content\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.539739 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-utilities\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.539775 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqrs7\" (UniqueName: \"kubernetes.io/projected/bf27d2cb-afca-4848-908f-c3d8c6fd7863-kube-api-access-sqrs7\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.641834 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-catalog-content\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.641884 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-utilities\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.641910 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqrs7\" (UniqueName: \"kubernetes.io/projected/bf27d2cb-afca-4848-908f-c3d8c6fd7863-kube-api-access-sqrs7\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.642331 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-catalog-content\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.642485 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-utilities\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.660975 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqrs7\" (UniqueName: \"kubernetes.io/projected/bf27d2cb-afca-4848-908f-c3d8c6fd7863-kube-api-access-sqrs7\") pod \"certified-operators-vzkls\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:37 crc kubenswrapper[4669]: I1010 09:49:37.797214 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:38 crc kubenswrapper[4669]: I1010 09:49:38.336687 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzkls"] Oct 10 09:49:38 crc kubenswrapper[4669]: I1010 09:49:38.690298 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerID="5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c" exitCode=0 Oct 10 09:49:38 crc kubenswrapper[4669]: I1010 09:49:38.690383 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzkls" event={"ID":"bf27d2cb-afca-4848-908f-c3d8c6fd7863","Type":"ContainerDied","Data":"5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c"} Oct 10 09:49:38 crc kubenswrapper[4669]: I1010 09:49:38.690639 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzkls" event={"ID":"bf27d2cb-afca-4848-908f-c3d8c6fd7863","Type":"ContainerStarted","Data":"d448baf8d9c2089b214c242dde578bcb24fa3524656bb091a1fc93dea2c6bf04"} Oct 10 09:49:39 crc kubenswrapper[4669]: I1010 09:49:39.701843 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzkls" event={"ID":"bf27d2cb-afca-4848-908f-c3d8c6fd7863","Type":"ContainerStarted","Data":"b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07"} Oct 10 09:49:40 crc kubenswrapper[4669]: I1010 09:49:40.710499 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerID="b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07" exitCode=0 Oct 10 09:49:40 crc kubenswrapper[4669]: I1010 09:49:40.710645 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzkls" event={"ID":"bf27d2cb-afca-4848-908f-c3d8c6fd7863","Type":"ContainerDied","Data":"b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07"} Oct 10 09:49:41 crc kubenswrapper[4669]: I1010 09:49:41.730929 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzkls" event={"ID":"bf27d2cb-afca-4848-908f-c3d8c6fd7863","Type":"ContainerStarted","Data":"5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa"} Oct 10 09:49:41 crc kubenswrapper[4669]: I1010 09:49:41.751087 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vzkls" podStartSLOduration=2.142479132 podStartE2EDuration="4.751064335s" podCreationTimestamp="2025-10-10 09:49:37 +0000 UTC" firstStartedPulling="2025-10-10 09:49:38.693018197 +0000 UTC m=+2321.709036939" lastFinishedPulling="2025-10-10 09:49:41.30160339 +0000 UTC m=+2324.317622142" observedRunningTime="2025-10-10 09:49:41.745242608 +0000 UTC m=+2324.761261370" watchObservedRunningTime="2025-10-10 09:49:41.751064335 +0000 UTC m=+2324.767083077" Oct 10 09:49:47 crc kubenswrapper[4669]: I1010 09:49:47.810021 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:47 crc kubenswrapper[4669]: I1010 09:49:47.810554 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:47 crc kubenswrapper[4669]: I1010 09:49:47.845338 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:48 crc kubenswrapper[4669]: I1010 09:49:48.862713 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:48 crc kubenswrapper[4669]: I1010 09:49:48.925927 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzkls"] Oct 10 09:49:50 crc kubenswrapper[4669]: I1010 09:49:50.794878 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:49:50 crc kubenswrapper[4669]: E1010 09:49:50.796039 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:49:50 crc kubenswrapper[4669]: I1010 09:49:50.809538 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vzkls" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="registry-server" containerID="cri-o://5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa" gracePeriod=2 Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.252744 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.305485 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqrs7\" (UniqueName: \"kubernetes.io/projected/bf27d2cb-afca-4848-908f-c3d8c6fd7863-kube-api-access-sqrs7\") pod \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.305650 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-catalog-content\") pod \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.305732 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-utilities\") pod \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\" (UID: \"bf27d2cb-afca-4848-908f-c3d8c6fd7863\") " Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.306606 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-utilities" (OuterVolumeSpecName: "utilities") pod "bf27d2cb-afca-4848-908f-c3d8c6fd7863" (UID: "bf27d2cb-afca-4848-908f-c3d8c6fd7863"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.317361 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf27d2cb-afca-4848-908f-c3d8c6fd7863-kube-api-access-sqrs7" (OuterVolumeSpecName: "kube-api-access-sqrs7") pod "bf27d2cb-afca-4848-908f-c3d8c6fd7863" (UID: "bf27d2cb-afca-4848-908f-c3d8c6fd7863"). InnerVolumeSpecName "kube-api-access-sqrs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.357961 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf27d2cb-afca-4848-908f-c3d8c6fd7863" (UID: "bf27d2cb-afca-4848-908f-c3d8c6fd7863"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.408208 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqrs7\" (UniqueName: \"kubernetes.io/projected/bf27d2cb-afca-4848-908f-c3d8c6fd7863-kube-api-access-sqrs7\") on node \"crc\" DevicePath \"\"" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.408539 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.408552 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf27d2cb-afca-4848-908f-c3d8c6fd7863-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.834600 4669 generic.go:334] "Generic (PLEG): container finished" podID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerID="5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa" exitCode=0 Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.834652 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzkls" event={"ID":"bf27d2cb-afca-4848-908f-c3d8c6fd7863","Type":"ContainerDied","Data":"5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa"} Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.834698 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzkls" event={"ID":"bf27d2cb-afca-4848-908f-c3d8c6fd7863","Type":"ContainerDied","Data":"d448baf8d9c2089b214c242dde578bcb24fa3524656bb091a1fc93dea2c6bf04"} Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.834719 4669 scope.go:117] "RemoveContainer" containerID="5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.834980 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzkls" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.863421 4669 scope.go:117] "RemoveContainer" containerID="b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.901844 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzkls"] Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.904261 4669 scope.go:117] "RemoveContainer" containerID="5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.909241 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vzkls"] Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.929056 4669 scope.go:117] "RemoveContainer" containerID="5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa" Oct 10 09:49:51 crc kubenswrapper[4669]: E1010 09:49:51.930037 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa\": container with ID starting with 5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa not found: ID does not exist" containerID="5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.930080 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa"} err="failed to get container status \"5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa\": rpc error: code = NotFound desc = could not find container \"5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa\": container with ID starting with 5127df6fbb7b8288722beb6613c9dfb08f7eb4264f4a7eb0b479e4f4232623fa not found: ID does not exist" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.930105 4669 scope.go:117] "RemoveContainer" containerID="b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07" Oct 10 09:49:51 crc kubenswrapper[4669]: E1010 09:49:51.930561 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07\": container with ID starting with b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07 not found: ID does not exist" containerID="b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.930619 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07"} err="failed to get container status \"b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07\": rpc error: code = NotFound desc = could not find container \"b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07\": container with ID starting with b437bd326029cbfe7ac86b17abb16d09bebcfb939f0774cbe61bcce5f4368d07 not found: ID does not exist" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.930644 4669 scope.go:117] "RemoveContainer" containerID="5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c" Oct 10 09:49:51 crc kubenswrapper[4669]: E1010 09:49:51.930953 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c\": container with ID starting with 5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c not found: ID does not exist" containerID="5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c" Oct 10 09:49:51 crc kubenswrapper[4669]: I1010 09:49:51.930972 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c"} err="failed to get container status \"5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c\": rpc error: code = NotFound desc = could not find container \"5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c\": container with ID starting with 5affd4cc21dd12d93ec1167da30e9003c10453e8bcd1b73d28d27534c0dcc97c not found: ID does not exist" Oct 10 09:49:53 crc kubenswrapper[4669]: I1010 09:49:53.807126 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" path="/var/lib/kubelet/pods/bf27d2cb-afca-4848-908f-c3d8c6fd7863/volumes" Oct 10 09:50:05 crc kubenswrapper[4669]: I1010 09:50:05.795762 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:50:05 crc kubenswrapper[4669]: E1010 09:50:05.796659 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:50:05 crc kubenswrapper[4669]: I1010 09:50:05.952743 4669 generic.go:334] "Generic (PLEG): container finished" podID="66dbecf2-f169-4c6c-b8be-efbab612ace8" containerID="b4a689b2ef12140a336b76e11541a7e15ba6adb45cd65510a30ccfe5da9648b9" exitCode=0 Oct 10 09:50:05 crc kubenswrapper[4669]: I1010 09:50:05.952808 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" event={"ID":"66dbecf2-f169-4c6c-b8be-efbab612ace8","Type":"ContainerDied","Data":"b4a689b2ef12140a336b76e11541a7e15ba6adb45cd65510a30ccfe5da9648b9"} Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.396508 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.444690 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ssh-key\") pod \"66dbecf2-f169-4c6c-b8be-efbab612ace8\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.444758 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ceph\") pod \"66dbecf2-f169-4c6c-b8be-efbab612ace8\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.444820 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-inventory\") pod \"66dbecf2-f169-4c6c-b8be-efbab612ace8\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.444927 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6zqm\" (UniqueName: \"kubernetes.io/projected/66dbecf2-f169-4c6c-b8be-efbab612ace8-kube-api-access-z6zqm\") pod \"66dbecf2-f169-4c6c-b8be-efbab612ace8\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.444969 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-bootstrap-combined-ca-bundle\") pod \"66dbecf2-f169-4c6c-b8be-efbab612ace8\" (UID: \"66dbecf2-f169-4c6c-b8be-efbab612ace8\") " Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.451809 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "66dbecf2-f169-4c6c-b8be-efbab612ace8" (UID: "66dbecf2-f169-4c6c-b8be-efbab612ace8"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.452081 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66dbecf2-f169-4c6c-b8be-efbab612ace8-kube-api-access-z6zqm" (OuterVolumeSpecName: "kube-api-access-z6zqm") pod "66dbecf2-f169-4c6c-b8be-efbab612ace8" (UID: "66dbecf2-f169-4c6c-b8be-efbab612ace8"). InnerVolumeSpecName "kube-api-access-z6zqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.452308 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ceph" (OuterVolumeSpecName: "ceph") pod "66dbecf2-f169-4c6c-b8be-efbab612ace8" (UID: "66dbecf2-f169-4c6c-b8be-efbab612ace8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.482860 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66dbecf2-f169-4c6c-b8be-efbab612ace8" (UID: "66dbecf2-f169-4c6c-b8be-efbab612ace8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.483295 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-inventory" (OuterVolumeSpecName: "inventory") pod "66dbecf2-f169-4c6c-b8be-efbab612ace8" (UID: "66dbecf2-f169-4c6c-b8be-efbab612ace8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.546963 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.547012 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.547023 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.547033 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6zqm\" (UniqueName: \"kubernetes.io/projected/66dbecf2-f169-4c6c-b8be-efbab612ace8-kube-api-access-z6zqm\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.547044 4669 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66dbecf2-f169-4c6c-b8be-efbab612ace8-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.972397 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" event={"ID":"66dbecf2-f169-4c6c-b8be-efbab612ace8","Type":"ContainerDied","Data":"c12f6791ac6499a07e679fb7e2a4556ba21d21c166a671627a8fc8a7fbd03bdb"} Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.972671 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c12f6791ac6499a07e679fb7e2a4556ba21d21c166a671627a8fc8a7fbd03bdb" Oct 10 09:50:07 crc kubenswrapper[4669]: I1010 09:50:07.972429 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.058018 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg"] Oct 10 09:50:08 crc kubenswrapper[4669]: E1010 09:50:08.058774 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dbecf2-f169-4c6c-b8be-efbab612ace8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.058839 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dbecf2-f169-4c6c-b8be-efbab612ace8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:08 crc kubenswrapper[4669]: E1010 09:50:08.058892 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="registry-server" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.058948 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="registry-server" Oct 10 09:50:08 crc kubenswrapper[4669]: E1010 09:50:08.059009 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="extract-content" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.059057 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="extract-content" Oct 10 09:50:08 crc kubenswrapper[4669]: E1010 09:50:08.059124 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="extract-utilities" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.059177 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="extract-utilities" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.059375 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="66dbecf2-f169-4c6c-b8be-efbab612ace8" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.059447 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf27d2cb-afca-4848-908f-c3d8c6fd7863" containerName="registry-server" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.060042 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.063846 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.064005 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.064508 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.065136 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.069024 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.079997 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg"] Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.159931 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.160008 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.160067 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7flt\" (UniqueName: \"kubernetes.io/projected/7469da7f-e608-439c-93b1-d60130d75a72-kube-api-access-x7flt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.160142 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.261734 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.262181 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.262823 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.262981 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7flt\" (UniqueName: \"kubernetes.io/projected/7469da7f-e608-439c-93b1-d60130d75a72-kube-api-access-x7flt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.267038 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.267161 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.273886 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.290004 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7flt\" (UniqueName: \"kubernetes.io/projected/7469da7f-e608-439c-93b1-d60130d75a72-kube-api-access-x7flt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.375654 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:08 crc kubenswrapper[4669]: I1010 09:50:08.986741 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg"] Oct 10 09:50:09 crc kubenswrapper[4669]: I1010 09:50:09.995073 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" event={"ID":"7469da7f-e608-439c-93b1-d60130d75a72","Type":"ContainerStarted","Data":"929077b20525a2f0445b522d7b0ce9b8204ba286c04c8ffd547ee3cb4f400671"} Oct 10 09:50:11 crc kubenswrapper[4669]: I1010 09:50:11.004265 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" event={"ID":"7469da7f-e608-439c-93b1-d60130d75a72","Type":"ContainerStarted","Data":"493cc345d0342d6e3806acc338efd4de92443589bbf9551187bc0c60932b8771"} Oct 10 09:50:11 crc kubenswrapper[4669]: I1010 09:50:11.022685 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" podStartSLOduration=2.466275429 podStartE2EDuration="3.022661661s" podCreationTimestamp="2025-10-10 09:50:08 +0000 UTC" firstStartedPulling="2025-10-10 09:50:08.996141418 +0000 UTC m=+2352.012160160" lastFinishedPulling="2025-10-10 09:50:09.55252764 +0000 UTC m=+2352.568546392" observedRunningTime="2025-10-10 09:50:11.019544522 +0000 UTC m=+2354.035563264" watchObservedRunningTime="2025-10-10 09:50:11.022661661 +0000 UTC m=+2354.038680423" Oct 10 09:50:18 crc kubenswrapper[4669]: I1010 09:50:18.795537 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:50:18 crc kubenswrapper[4669]: E1010 09:50:18.796401 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:50:29 crc kubenswrapper[4669]: I1010 09:50:29.795854 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:50:29 crc kubenswrapper[4669]: E1010 09:50:29.796832 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:50:38 crc kubenswrapper[4669]: I1010 09:50:38.212205 4669 generic.go:334] "Generic (PLEG): container finished" podID="7469da7f-e608-439c-93b1-d60130d75a72" containerID="493cc345d0342d6e3806acc338efd4de92443589bbf9551187bc0c60932b8771" exitCode=0 Oct 10 09:50:38 crc kubenswrapper[4669]: I1010 09:50:38.212305 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" event={"ID":"7469da7f-e608-439c-93b1-d60130d75a72","Type":"ContainerDied","Data":"493cc345d0342d6e3806acc338efd4de92443589bbf9551187bc0c60932b8771"} Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.670612 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.793176 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-inventory\") pod \"7469da7f-e608-439c-93b1-d60130d75a72\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.793246 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ceph\") pod \"7469da7f-e608-439c-93b1-d60130d75a72\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.793367 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7flt\" (UniqueName: \"kubernetes.io/projected/7469da7f-e608-439c-93b1-d60130d75a72-kube-api-access-x7flt\") pod \"7469da7f-e608-439c-93b1-d60130d75a72\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.793406 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ssh-key\") pod \"7469da7f-e608-439c-93b1-d60130d75a72\" (UID: \"7469da7f-e608-439c-93b1-d60130d75a72\") " Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.802708 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ceph" (OuterVolumeSpecName: "ceph") pod "7469da7f-e608-439c-93b1-d60130d75a72" (UID: "7469da7f-e608-439c-93b1-d60130d75a72"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.858871 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7469da7f-e608-439c-93b1-d60130d75a72-kube-api-access-x7flt" (OuterVolumeSpecName: "kube-api-access-x7flt") pod "7469da7f-e608-439c-93b1-d60130d75a72" (UID: "7469da7f-e608-439c-93b1-d60130d75a72"). InnerVolumeSpecName "kube-api-access-x7flt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.896471 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.896603 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7flt\" (UniqueName: \"kubernetes.io/projected/7469da7f-e608-439c-93b1-d60130d75a72-kube-api-access-x7flt\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.953116 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-inventory" (OuterVolumeSpecName: "inventory") pod "7469da7f-e608-439c-93b1-d60130d75a72" (UID: "7469da7f-e608-439c-93b1-d60130d75a72"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.959804 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7469da7f-e608-439c-93b1-d60130d75a72" (UID: "7469da7f-e608-439c-93b1-d60130d75a72"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.998599 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:39 crc kubenswrapper[4669]: I1010 09:50:39.998638 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7469da7f-e608-439c-93b1-d60130d75a72-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.227959 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" event={"ID":"7469da7f-e608-439c-93b1-d60130d75a72","Type":"ContainerDied","Data":"929077b20525a2f0445b522d7b0ce9b8204ba286c04c8ffd547ee3cb4f400671"} Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.228001 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="929077b20525a2f0445b522d7b0ce9b8204ba286c04c8ffd547ee3cb4f400671" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.228052 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.329045 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl"] Oct 10 09:50:40 crc kubenswrapper[4669]: E1010 09:50:40.329530 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7469da7f-e608-439c-93b1-d60130d75a72" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.329549 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7469da7f-e608-439c-93b1-d60130d75a72" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.329769 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7469da7f-e608-439c-93b1-d60130d75a72" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.330335 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.332922 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.333470 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.334142 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.334232 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.338835 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.352143 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl"] Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.507746 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.507838 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.507919 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7xdh\" (UniqueName: \"kubernetes.io/projected/4feb106b-763c-422d-a1df-a95b354a7770-kube-api-access-n7xdh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.507956 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.609397 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.609512 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.609558 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.609625 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7xdh\" (UniqueName: \"kubernetes.io/projected/4feb106b-763c-422d-a1df-a95b354a7770-kube-api-access-n7xdh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.614051 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.614082 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.614193 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.626887 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7xdh\" (UniqueName: \"kubernetes.io/projected/4feb106b-763c-422d-a1df-a95b354a7770-kube-api-access-n7xdh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:40 crc kubenswrapper[4669]: I1010 09:50:40.649855 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:41 crc kubenswrapper[4669]: I1010 09:50:41.206601 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl"] Oct 10 09:50:41 crc kubenswrapper[4669]: I1010 09:50:41.245696 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" event={"ID":"4feb106b-763c-422d-a1df-a95b354a7770","Type":"ContainerStarted","Data":"0f95a8e5ac253025886e0c99baa260c3cc6860ec471a5c284a878196dfbfa4e1"} Oct 10 09:50:42 crc kubenswrapper[4669]: I1010 09:50:42.257185 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" event={"ID":"4feb106b-763c-422d-a1df-a95b354a7770","Type":"ContainerStarted","Data":"279aabd69cf765a657155505b67a47d9513c7726ab598e890ba8e44223e86cb6"} Oct 10 09:50:44 crc kubenswrapper[4669]: I1010 09:50:44.795410 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:50:44 crc kubenswrapper[4669]: E1010 09:50:44.795983 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:50:48 crc kubenswrapper[4669]: I1010 09:50:48.312612 4669 generic.go:334] "Generic (PLEG): container finished" podID="4feb106b-763c-422d-a1df-a95b354a7770" containerID="279aabd69cf765a657155505b67a47d9513c7726ab598e890ba8e44223e86cb6" exitCode=0 Oct 10 09:50:48 crc kubenswrapper[4669]: I1010 09:50:48.312684 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" event={"ID":"4feb106b-763c-422d-a1df-a95b354a7770","Type":"ContainerDied","Data":"279aabd69cf765a657155505b67a47d9513c7726ab598e890ba8e44223e86cb6"} Oct 10 09:50:49 crc kubenswrapper[4669]: I1010 09:50:49.810840 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:49 crc kubenswrapper[4669]: I1010 09:50:49.984112 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7xdh\" (UniqueName: \"kubernetes.io/projected/4feb106b-763c-422d-a1df-a95b354a7770-kube-api-access-n7xdh\") pod \"4feb106b-763c-422d-a1df-a95b354a7770\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " Oct 10 09:50:49 crc kubenswrapper[4669]: I1010 09:50:49.984186 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ceph\") pod \"4feb106b-763c-422d-a1df-a95b354a7770\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " Oct 10 09:50:49 crc kubenswrapper[4669]: I1010 09:50:49.984222 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ssh-key\") pod \"4feb106b-763c-422d-a1df-a95b354a7770\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " Oct 10 09:50:49 crc kubenswrapper[4669]: I1010 09:50:49.984437 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-inventory\") pod \"4feb106b-763c-422d-a1df-a95b354a7770\" (UID: \"4feb106b-763c-422d-a1df-a95b354a7770\") " Oct 10 09:50:49 crc kubenswrapper[4669]: I1010 09:50:49.990190 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ceph" (OuterVolumeSpecName: "ceph") pod "4feb106b-763c-422d-a1df-a95b354a7770" (UID: "4feb106b-763c-422d-a1df-a95b354a7770"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:49 crc kubenswrapper[4669]: I1010 09:50:49.990496 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4feb106b-763c-422d-a1df-a95b354a7770-kube-api-access-n7xdh" (OuterVolumeSpecName: "kube-api-access-n7xdh") pod "4feb106b-763c-422d-a1df-a95b354a7770" (UID: "4feb106b-763c-422d-a1df-a95b354a7770"). InnerVolumeSpecName "kube-api-access-n7xdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.019105 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4feb106b-763c-422d-a1df-a95b354a7770" (UID: "4feb106b-763c-422d-a1df-a95b354a7770"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.020071 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-inventory" (OuterVolumeSpecName: "inventory") pod "4feb106b-763c-422d-a1df-a95b354a7770" (UID: "4feb106b-763c-422d-a1df-a95b354a7770"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.086575 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.086951 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7xdh\" (UniqueName: \"kubernetes.io/projected/4feb106b-763c-422d-a1df-a95b354a7770-kube-api-access-n7xdh\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.086965 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.086980 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4feb106b-763c-422d-a1df-a95b354a7770-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.331221 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" event={"ID":"4feb106b-763c-422d-a1df-a95b354a7770","Type":"ContainerDied","Data":"0f95a8e5ac253025886e0c99baa260c3cc6860ec471a5c284a878196dfbfa4e1"} Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.331258 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f95a8e5ac253025886e0c99baa260c3cc6860ec471a5c284a878196dfbfa4e1" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.331296 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.426059 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz"] Oct 10 09:50:50 crc kubenswrapper[4669]: E1010 09:50:50.426454 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4feb106b-763c-422d-a1df-a95b354a7770" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.426476 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="4feb106b-763c-422d-a1df-a95b354a7770" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.426719 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="4feb106b-763c-422d-a1df-a95b354a7770" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.427449 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.429994 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.430006 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.430196 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.430473 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.432709 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.434447 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz"] Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.595714 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clggf\" (UniqueName: \"kubernetes.io/projected/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-kube-api-access-clggf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.595997 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.596070 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.596242 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.715742 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.715793 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.715832 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.715862 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clggf\" (UniqueName: \"kubernetes.io/projected/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-kube-api-access-clggf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.722224 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.722848 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.724416 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.750118 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clggf\" (UniqueName: \"kubernetes.io/projected/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-kube-api-access-clggf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hzlnz\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:50 crc kubenswrapper[4669]: I1010 09:50:50.752810 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:50:51 crc kubenswrapper[4669]: I1010 09:50:51.312638 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz"] Oct 10 09:50:51 crc kubenswrapper[4669]: I1010 09:50:51.340056 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" event={"ID":"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f","Type":"ContainerStarted","Data":"5c210ccb944c79d151e9f707bea216c0c3451d099b7befa177a68aedb8ed554a"} Oct 10 09:50:52 crc kubenswrapper[4669]: I1010 09:50:52.349850 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" event={"ID":"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f","Type":"ContainerStarted","Data":"7f7e0cf47a1df510ef2121381860cc3c6d01d3bde7ac677e966d110bf7190988"} Oct 10 09:50:52 crc kubenswrapper[4669]: I1010 09:50:52.368768 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" podStartSLOduration=1.7951682500000001 podStartE2EDuration="2.368731112s" podCreationTimestamp="2025-10-10 09:50:50 +0000 UTC" firstStartedPulling="2025-10-10 09:50:51.32211137 +0000 UTC m=+2394.338130112" lastFinishedPulling="2025-10-10 09:50:51.895674232 +0000 UTC m=+2394.911692974" observedRunningTime="2025-10-10 09:50:52.362166191 +0000 UTC m=+2395.378184933" watchObservedRunningTime="2025-10-10 09:50:52.368731112 +0000 UTC m=+2395.384749854" Oct 10 09:50:58 crc kubenswrapper[4669]: I1010 09:50:58.796091 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:50:58 crc kubenswrapper[4669]: E1010 09:50:58.796808 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:51:12 crc kubenswrapper[4669]: I1010 09:51:12.796642 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:51:12 crc kubenswrapper[4669]: E1010 09:51:12.797259 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:51:23 crc kubenswrapper[4669]: I1010 09:51:23.796490 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:51:23 crc kubenswrapper[4669]: E1010 09:51:23.797324 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:51:33 crc kubenswrapper[4669]: I1010 09:51:33.732523 4669 generic.go:334] "Generic (PLEG): container finished" podID="9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" containerID="7f7e0cf47a1df510ef2121381860cc3c6d01d3bde7ac677e966d110bf7190988" exitCode=0 Oct 10 09:51:33 crc kubenswrapper[4669]: I1010 09:51:33.732624 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" event={"ID":"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f","Type":"ContainerDied","Data":"7f7e0cf47a1df510ef2121381860cc3c6d01d3bde7ac677e966d110bf7190988"} Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.120704 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.162326 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ceph\") pod \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.162426 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-inventory\") pod \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.162475 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ssh-key\") pod \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.162503 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clggf\" (UniqueName: \"kubernetes.io/projected/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-kube-api-access-clggf\") pod \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\" (UID: \"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f\") " Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.167754 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-kube-api-access-clggf" (OuterVolumeSpecName: "kube-api-access-clggf") pod "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" (UID: "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f"). InnerVolumeSpecName "kube-api-access-clggf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.168455 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ceph" (OuterVolumeSpecName: "ceph") pod "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" (UID: "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.186861 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" (UID: "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.193445 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-inventory" (OuterVolumeSpecName: "inventory") pod "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" (UID: "9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.264920 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.265221 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.265233 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.265243 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clggf\" (UniqueName: \"kubernetes.io/projected/9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f-kube-api-access-clggf\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.748907 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" event={"ID":"9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f","Type":"ContainerDied","Data":"5c210ccb944c79d151e9f707bea216c0c3451d099b7befa177a68aedb8ed554a"} Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.748941 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hzlnz" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.748952 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c210ccb944c79d151e9f707bea216c0c3451d099b7befa177a68aedb8ed554a" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.922937 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv"] Oct 10 09:51:35 crc kubenswrapper[4669]: E1010 09:51:35.923756 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.923876 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.924182 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.925098 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.928443 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.928845 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.929020 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.929651 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.929957 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.947056 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv"] Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.979486 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.980018 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.980142 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:35 crc kubenswrapper[4669]: I1010 09:51:35.980287 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dfg9\" (UniqueName: \"kubernetes.io/projected/e53d40a4-c456-4991-8cfc-4dc303b5205d-kube-api-access-7dfg9\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.081258 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.081372 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.081427 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.082644 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dfg9\" (UniqueName: \"kubernetes.io/projected/e53d40a4-c456-4991-8cfc-4dc303b5205d-kube-api-access-7dfg9\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.086134 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.087997 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.091710 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.109867 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dfg9\" (UniqueName: \"kubernetes.io/projected/e53d40a4-c456-4991-8cfc-4dc303b5205d-kube-api-access-7dfg9\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.247187 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:36 crc kubenswrapper[4669]: I1010 09:51:36.803644 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv"] Oct 10 09:51:37 crc kubenswrapper[4669]: I1010 09:51:37.766478 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" event={"ID":"e53d40a4-c456-4991-8cfc-4dc303b5205d","Type":"ContainerStarted","Data":"779c0232c678b6b3a8b41fd22078e864c6996a284f0bb04b0ae2f71e52bfd5e0"} Oct 10 09:51:37 crc kubenswrapper[4669]: I1010 09:51:37.766972 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" event={"ID":"e53d40a4-c456-4991-8cfc-4dc303b5205d","Type":"ContainerStarted","Data":"3bcdc9057d6243f4750fecad8ddca9cfef86725bda0ec469069e469ad8178e52"} Oct 10 09:51:37 crc kubenswrapper[4669]: I1010 09:51:37.789357 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" podStartSLOduration=2.220420274 podStartE2EDuration="2.789339577s" podCreationTimestamp="2025-10-10 09:51:35 +0000 UTC" firstStartedPulling="2025-10-10 09:51:36.813759994 +0000 UTC m=+2439.829778736" lastFinishedPulling="2025-10-10 09:51:37.382679277 +0000 UTC m=+2440.398698039" observedRunningTime="2025-10-10 09:51:37.785287177 +0000 UTC m=+2440.801305919" watchObservedRunningTime="2025-10-10 09:51:37.789339577 +0000 UTC m=+2440.805358319" Oct 10 09:51:38 crc kubenswrapper[4669]: I1010 09:51:38.796329 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:51:38 crc kubenswrapper[4669]: E1010 09:51:38.796851 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:51:41 crc kubenswrapper[4669]: I1010 09:51:41.800387 4669 generic.go:334] "Generic (PLEG): container finished" podID="e53d40a4-c456-4991-8cfc-4dc303b5205d" containerID="779c0232c678b6b3a8b41fd22078e864c6996a284f0bb04b0ae2f71e52bfd5e0" exitCode=0 Oct 10 09:51:41 crc kubenswrapper[4669]: I1010 09:51:41.808008 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" event={"ID":"e53d40a4-c456-4991-8cfc-4dc303b5205d","Type":"ContainerDied","Data":"779c0232c678b6b3a8b41fd22078e864c6996a284f0bb04b0ae2f71e52bfd5e0"} Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.237510 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.430374 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ssh-key\") pod \"e53d40a4-c456-4991-8cfc-4dc303b5205d\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.430527 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dfg9\" (UniqueName: \"kubernetes.io/projected/e53d40a4-c456-4991-8cfc-4dc303b5205d-kube-api-access-7dfg9\") pod \"e53d40a4-c456-4991-8cfc-4dc303b5205d\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.430558 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-inventory\") pod \"e53d40a4-c456-4991-8cfc-4dc303b5205d\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.430729 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ceph\") pod \"e53d40a4-c456-4991-8cfc-4dc303b5205d\" (UID: \"e53d40a4-c456-4991-8cfc-4dc303b5205d\") " Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.441905 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ceph" (OuterVolumeSpecName: "ceph") pod "e53d40a4-c456-4991-8cfc-4dc303b5205d" (UID: "e53d40a4-c456-4991-8cfc-4dc303b5205d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.449820 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e53d40a4-c456-4991-8cfc-4dc303b5205d-kube-api-access-7dfg9" (OuterVolumeSpecName: "kube-api-access-7dfg9") pod "e53d40a4-c456-4991-8cfc-4dc303b5205d" (UID: "e53d40a4-c456-4991-8cfc-4dc303b5205d"). InnerVolumeSpecName "kube-api-access-7dfg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.460269 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e53d40a4-c456-4991-8cfc-4dc303b5205d" (UID: "e53d40a4-c456-4991-8cfc-4dc303b5205d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.473404 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-inventory" (OuterVolumeSpecName: "inventory") pod "e53d40a4-c456-4991-8cfc-4dc303b5205d" (UID: "e53d40a4-c456-4991-8cfc-4dc303b5205d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.532873 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.533080 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.533204 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dfg9\" (UniqueName: \"kubernetes.io/projected/e53d40a4-c456-4991-8cfc-4dc303b5205d-kube-api-access-7dfg9\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.533287 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53d40a4-c456-4991-8cfc-4dc303b5205d-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.817699 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" event={"ID":"e53d40a4-c456-4991-8cfc-4dc303b5205d","Type":"ContainerDied","Data":"3bcdc9057d6243f4750fecad8ddca9cfef86725bda0ec469069e469ad8178e52"} Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.817762 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bcdc9057d6243f4750fecad8ddca9cfef86725bda0ec469069e469ad8178e52" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.817948 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.925607 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8"] Oct 10 09:51:43 crc kubenswrapper[4669]: E1010 09:51:43.926008 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53d40a4-c456-4991-8cfc-4dc303b5205d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.926026 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53d40a4-c456-4991-8cfc-4dc303b5205d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.926216 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53d40a4-c456-4991-8cfc-4dc303b5205d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.926788 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.928211 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.928547 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.928927 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.930329 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.930560 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:51:43 crc kubenswrapper[4669]: I1010 09:51:43.980209 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8"] Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.046679 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.046720 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb45f\" (UniqueName: \"kubernetes.io/projected/2d5855cf-c66a-4daa-9b94-84636ee8936d-kube-api-access-pb45f\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.046755 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.046821 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.147896 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.147940 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb45f\" (UniqueName: \"kubernetes.io/projected/2d5855cf-c66a-4daa-9b94-84636ee8936d-kube-api-access-pb45f\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.147971 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.148004 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.151040 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.151310 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.152337 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.171305 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb45f\" (UniqueName: \"kubernetes.io/projected/2d5855cf-c66a-4daa-9b94-84636ee8936d-kube-api-access-pb45f\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-st5n8\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.248801 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.779248 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8"] Oct 10 09:51:44 crc kubenswrapper[4669]: I1010 09:51:44.826990 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" event={"ID":"2d5855cf-c66a-4daa-9b94-84636ee8936d","Type":"ContainerStarted","Data":"ac15556b16d5a9154c73c34763e6b146c33564b782297515e14b06c80f1b2892"} Oct 10 09:51:45 crc kubenswrapper[4669]: I1010 09:51:45.838680 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" event={"ID":"2d5855cf-c66a-4daa-9b94-84636ee8936d","Type":"ContainerStarted","Data":"141f2ce1779c1dcf9d0f824a718085df4ba7152cecfd25917f201102359984e2"} Oct 10 09:51:45 crc kubenswrapper[4669]: I1010 09:51:45.857785 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" podStartSLOduration=2.38271555 podStartE2EDuration="2.857766643s" podCreationTimestamp="2025-10-10 09:51:43 +0000 UTC" firstStartedPulling="2025-10-10 09:51:44.792689749 +0000 UTC m=+2447.808708491" lastFinishedPulling="2025-10-10 09:51:45.267740842 +0000 UTC m=+2448.283759584" observedRunningTime="2025-10-10 09:51:45.856404339 +0000 UTC m=+2448.872423071" watchObservedRunningTime="2025-10-10 09:51:45.857766643 +0000 UTC m=+2448.873785385" Oct 10 09:51:53 crc kubenswrapper[4669]: I1010 09:51:53.796329 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:51:53 crc kubenswrapper[4669]: E1010 09:51:53.797544 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:52:07 crc kubenswrapper[4669]: I1010 09:52:07.804660 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:52:07 crc kubenswrapper[4669]: E1010 09:52:07.805615 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.564561 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xzkk9"] Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.576735 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.592949 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xzkk9"] Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.625469 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-utilities\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.626403 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-499z2\" (UniqueName: \"kubernetes.io/projected/1044dcfd-4670-4809-a829-d08f2dfb6629-kube-api-access-499z2\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.627308 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-catalog-content\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.728942 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-499z2\" (UniqueName: \"kubernetes.io/projected/1044dcfd-4670-4809-a829-d08f2dfb6629-kube-api-access-499z2\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.729063 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-catalog-content\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.729175 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-utilities\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.729568 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-catalog-content\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.729691 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-utilities\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.752467 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-499z2\" (UniqueName: \"kubernetes.io/projected/1044dcfd-4670-4809-a829-d08f2dfb6629-kube-api-access-499z2\") pod \"redhat-operators-xzkk9\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:14 crc kubenswrapper[4669]: I1010 09:52:14.910385 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:15 crc kubenswrapper[4669]: I1010 09:52:15.396824 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xzkk9"] Oct 10 09:52:16 crc kubenswrapper[4669]: I1010 09:52:16.095953 4669 generic.go:334] "Generic (PLEG): container finished" podID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerID="e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d" exitCode=0 Oct 10 09:52:16 crc kubenswrapper[4669]: I1010 09:52:16.096072 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzkk9" event={"ID":"1044dcfd-4670-4809-a829-d08f2dfb6629","Type":"ContainerDied","Data":"e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d"} Oct 10 09:52:16 crc kubenswrapper[4669]: I1010 09:52:16.096323 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzkk9" event={"ID":"1044dcfd-4670-4809-a829-d08f2dfb6629","Type":"ContainerStarted","Data":"4109f01e7311a73b4183838f562081661eea0721b6a4980e6d9fb353f3aeb862"} Oct 10 09:52:17 crc kubenswrapper[4669]: I1010 09:52:17.105735 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzkk9" event={"ID":"1044dcfd-4670-4809-a829-d08f2dfb6629","Type":"ContainerStarted","Data":"328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1"} Oct 10 09:52:18 crc kubenswrapper[4669]: I1010 09:52:18.117039 4669 generic.go:334] "Generic (PLEG): container finished" podID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerID="328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1" exitCode=0 Oct 10 09:52:18 crc kubenswrapper[4669]: I1010 09:52:18.117116 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzkk9" event={"ID":"1044dcfd-4670-4809-a829-d08f2dfb6629","Type":"ContainerDied","Data":"328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1"} Oct 10 09:52:19 crc kubenswrapper[4669]: I1010 09:52:19.131635 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzkk9" event={"ID":"1044dcfd-4670-4809-a829-d08f2dfb6629","Type":"ContainerStarted","Data":"0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38"} Oct 10 09:52:19 crc kubenswrapper[4669]: I1010 09:52:19.153662 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xzkk9" podStartSLOduration=2.506902149 podStartE2EDuration="5.153644136s" podCreationTimestamp="2025-10-10 09:52:14 +0000 UTC" firstStartedPulling="2025-10-10 09:52:16.099528053 +0000 UTC m=+2479.115546805" lastFinishedPulling="2025-10-10 09:52:18.74627005 +0000 UTC m=+2481.762288792" observedRunningTime="2025-10-10 09:52:19.149613008 +0000 UTC m=+2482.165631760" watchObservedRunningTime="2025-10-10 09:52:19.153644136 +0000 UTC m=+2482.169662868" Oct 10 09:52:22 crc kubenswrapper[4669]: I1010 09:52:22.795748 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:52:22 crc kubenswrapper[4669]: E1010 09:52:22.796627 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:52:24 crc kubenswrapper[4669]: I1010 09:52:24.911963 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:24 crc kubenswrapper[4669]: I1010 09:52:24.916069 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:24 crc kubenswrapper[4669]: I1010 09:52:24.956767 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:25 crc kubenswrapper[4669]: I1010 09:52:25.226936 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:25 crc kubenswrapper[4669]: I1010 09:52:25.304754 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xzkk9"] Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.193990 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xzkk9" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="registry-server" containerID="cri-o://0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38" gracePeriod=2 Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.684391 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.785696 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-catalog-content\") pod \"1044dcfd-4670-4809-a829-d08f2dfb6629\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.785866 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-499z2\" (UniqueName: \"kubernetes.io/projected/1044dcfd-4670-4809-a829-d08f2dfb6629-kube-api-access-499z2\") pod \"1044dcfd-4670-4809-a829-d08f2dfb6629\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.785944 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-utilities\") pod \"1044dcfd-4670-4809-a829-d08f2dfb6629\" (UID: \"1044dcfd-4670-4809-a829-d08f2dfb6629\") " Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.786853 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-utilities" (OuterVolumeSpecName: "utilities") pod "1044dcfd-4670-4809-a829-d08f2dfb6629" (UID: "1044dcfd-4670-4809-a829-d08f2dfb6629"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.792953 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1044dcfd-4670-4809-a829-d08f2dfb6629-kube-api-access-499z2" (OuterVolumeSpecName: "kube-api-access-499z2") pod "1044dcfd-4670-4809-a829-d08f2dfb6629" (UID: "1044dcfd-4670-4809-a829-d08f2dfb6629"). InnerVolumeSpecName "kube-api-access-499z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.876048 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1044dcfd-4670-4809-a829-d08f2dfb6629" (UID: "1044dcfd-4670-4809-a829-d08f2dfb6629"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.888745 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.888796 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1044dcfd-4670-4809-a829-d08f2dfb6629-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:27 crc kubenswrapper[4669]: I1010 09:52:27.888809 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-499z2\" (UniqueName: \"kubernetes.io/projected/1044dcfd-4670-4809-a829-d08f2dfb6629-kube-api-access-499z2\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.206028 4669 generic.go:334] "Generic (PLEG): container finished" podID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerID="0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38" exitCode=0 Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.206077 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzkk9" event={"ID":"1044dcfd-4670-4809-a829-d08f2dfb6629","Type":"ContainerDied","Data":"0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38"} Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.206132 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzkk9" event={"ID":"1044dcfd-4670-4809-a829-d08f2dfb6629","Type":"ContainerDied","Data":"4109f01e7311a73b4183838f562081661eea0721b6a4980e6d9fb353f3aeb862"} Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.206155 4669 scope.go:117] "RemoveContainer" containerID="0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.206157 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzkk9" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.230069 4669 scope.go:117] "RemoveContainer" containerID="328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.255002 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xzkk9"] Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.255848 4669 scope.go:117] "RemoveContainer" containerID="e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.262790 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xzkk9"] Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.302664 4669 scope.go:117] "RemoveContainer" containerID="0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38" Oct 10 09:52:28 crc kubenswrapper[4669]: E1010 09:52:28.303571 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38\": container with ID starting with 0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38 not found: ID does not exist" containerID="0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.303661 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38"} err="failed to get container status \"0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38\": rpc error: code = NotFound desc = could not find container \"0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38\": container with ID starting with 0feb2b416ef0c93d2a9c30b001c3c225375b1620c2ec0219ab25bc37bef75f38 not found: ID does not exist" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.303687 4669 scope.go:117] "RemoveContainer" containerID="328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1" Oct 10 09:52:28 crc kubenswrapper[4669]: E1010 09:52:28.304048 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1\": container with ID starting with 328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1 not found: ID does not exist" containerID="328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.304098 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1"} err="failed to get container status \"328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1\": rpc error: code = NotFound desc = could not find container \"328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1\": container with ID starting with 328041d6c72925d901923cbc90c4faf7637fc61bac1b8d637907a07d8789a0f1 not found: ID does not exist" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.304129 4669 scope.go:117] "RemoveContainer" containerID="e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d" Oct 10 09:52:28 crc kubenswrapper[4669]: E1010 09:52:28.304430 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d\": container with ID starting with e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d not found: ID does not exist" containerID="e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d" Oct 10 09:52:28 crc kubenswrapper[4669]: I1010 09:52:28.304458 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d"} err="failed to get container status \"e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d\": rpc error: code = NotFound desc = could not find container \"e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d\": container with ID starting with e592b75e15c4ecba8afe0a3d14ef4030f624dc9fb259f0a78cd116202a9c106d not found: ID does not exist" Oct 10 09:52:29 crc kubenswrapper[4669]: I1010 09:52:29.821692 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" path="/var/lib/kubelet/pods/1044dcfd-4670-4809-a829-d08f2dfb6629/volumes" Oct 10 09:52:35 crc kubenswrapper[4669]: I1010 09:52:35.284440 4669 generic.go:334] "Generic (PLEG): container finished" podID="2d5855cf-c66a-4daa-9b94-84636ee8936d" containerID="141f2ce1779c1dcf9d0f824a718085df4ba7152cecfd25917f201102359984e2" exitCode=0 Oct 10 09:52:35 crc kubenswrapper[4669]: I1010 09:52:35.284823 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" event={"ID":"2d5855cf-c66a-4daa-9b94-84636ee8936d","Type":"ContainerDied","Data":"141f2ce1779c1dcf9d0f824a718085df4ba7152cecfd25917f201102359984e2"} Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.717567 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.763874 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ssh-key\") pod \"2d5855cf-c66a-4daa-9b94-84636ee8936d\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.763920 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ceph\") pod \"2d5855cf-c66a-4daa-9b94-84636ee8936d\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.763966 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb45f\" (UniqueName: \"kubernetes.io/projected/2d5855cf-c66a-4daa-9b94-84636ee8936d-kube-api-access-pb45f\") pod \"2d5855cf-c66a-4daa-9b94-84636ee8936d\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.763999 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-inventory\") pod \"2d5855cf-c66a-4daa-9b94-84636ee8936d\" (UID: \"2d5855cf-c66a-4daa-9b94-84636ee8936d\") " Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.771413 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ceph" (OuterVolumeSpecName: "ceph") pod "2d5855cf-c66a-4daa-9b94-84636ee8936d" (UID: "2d5855cf-c66a-4daa-9b94-84636ee8936d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.776341 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d5855cf-c66a-4daa-9b94-84636ee8936d-kube-api-access-pb45f" (OuterVolumeSpecName: "kube-api-access-pb45f") pod "2d5855cf-c66a-4daa-9b94-84636ee8936d" (UID: "2d5855cf-c66a-4daa-9b94-84636ee8936d"). InnerVolumeSpecName "kube-api-access-pb45f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.792032 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2d5855cf-c66a-4daa-9b94-84636ee8936d" (UID: "2d5855cf-c66a-4daa-9b94-84636ee8936d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.796624 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-inventory" (OuterVolumeSpecName: "inventory") pod "2d5855cf-c66a-4daa-9b94-84636ee8936d" (UID: "2d5855cf-c66a-4daa-9b94-84636ee8936d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.865829 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb45f\" (UniqueName: \"kubernetes.io/projected/2d5855cf-c66a-4daa-9b94-84636ee8936d-kube-api-access-pb45f\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.865881 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.865896 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:36 crc kubenswrapper[4669]: I1010 09:52:36.865907 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2d5855cf-c66a-4daa-9b94-84636ee8936d-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.303540 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" event={"ID":"2d5855cf-c66a-4daa-9b94-84636ee8936d","Type":"ContainerDied","Data":"ac15556b16d5a9154c73c34763e6b146c33564b782297515e14b06c80f1b2892"} Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.303832 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac15556b16d5a9154c73c34763e6b146c33564b782297515e14b06c80f1b2892" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.303675 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-st5n8" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.407979 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dpnrk"] Oct 10 09:52:37 crc kubenswrapper[4669]: E1010 09:52:37.408538 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="extract-utilities" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.408568 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="extract-utilities" Oct 10 09:52:37 crc kubenswrapper[4669]: E1010 09:52:37.408638 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="registry-server" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.408657 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="registry-server" Oct 10 09:52:37 crc kubenswrapper[4669]: E1010 09:52:37.408689 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="extract-content" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.408707 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="extract-content" Oct 10 09:52:37 crc kubenswrapper[4669]: E1010 09:52:37.408749 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d5855cf-c66a-4daa-9b94-84636ee8936d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.408769 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d5855cf-c66a-4daa-9b94-84636ee8936d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.409073 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d5855cf-c66a-4daa-9b94-84636ee8936d" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.409120 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="1044dcfd-4670-4809-a829-d08f2dfb6629" containerName="registry-server" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.411120 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.419837 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.420000 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.420944 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.421089 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.421538 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.431196 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dpnrk"] Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.475215 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gch7\" (UniqueName: \"kubernetes.io/projected/5dc32d38-3a9b-48bb-93c4-76f51d215213-kube-api-access-2gch7\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.475359 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.475410 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.475451 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ceph\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.576994 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gch7\" (UniqueName: \"kubernetes.io/projected/5dc32d38-3a9b-48bb-93c4-76f51d215213-kube-api-access-2gch7\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.577648 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.578645 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.578767 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ceph\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.582492 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.583624 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.592566 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ceph\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.599788 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gch7\" (UniqueName: \"kubernetes.io/projected/5dc32d38-3a9b-48bb-93c4-76f51d215213-kube-api-access-2gch7\") pod \"ssh-known-hosts-edpm-deployment-dpnrk\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.726769 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:37 crc kubenswrapper[4669]: I1010 09:52:37.803194 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:52:37 crc kubenswrapper[4669]: E1010 09:52:37.803603 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:52:38 crc kubenswrapper[4669]: I1010 09:52:38.341420 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dpnrk"] Oct 10 09:52:39 crc kubenswrapper[4669]: I1010 09:52:39.325280 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" event={"ID":"5dc32d38-3a9b-48bb-93c4-76f51d215213","Type":"ContainerStarted","Data":"a5d90cb8e32cc650179cace4a019a84dc29edc8bd7ae0b9dd7293bfa4157bdab"} Oct 10 09:52:39 crc kubenswrapper[4669]: I1010 09:52:39.325550 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" event={"ID":"5dc32d38-3a9b-48bb-93c4-76f51d215213","Type":"ContainerStarted","Data":"47d49bbb62b4ab8d9ca9f63329b156c1b8582234616159f4dcf661e1ad3a9b28"} Oct 10 09:52:39 crc kubenswrapper[4669]: I1010 09:52:39.353967 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" podStartSLOduration=1.826777598 podStartE2EDuration="2.353949518s" podCreationTimestamp="2025-10-10 09:52:37 +0000 UTC" firstStartedPulling="2025-10-10 09:52:38.345971996 +0000 UTC m=+2501.361990778" lastFinishedPulling="2025-10-10 09:52:38.873143966 +0000 UTC m=+2501.889162698" observedRunningTime="2025-10-10 09:52:39.346927552 +0000 UTC m=+2502.362946324" watchObservedRunningTime="2025-10-10 09:52:39.353949518 +0000 UTC m=+2502.369968260" Oct 10 09:52:49 crc kubenswrapper[4669]: I1010 09:52:49.402164 4669 generic.go:334] "Generic (PLEG): container finished" podID="5dc32d38-3a9b-48bb-93c4-76f51d215213" containerID="a5d90cb8e32cc650179cace4a019a84dc29edc8bd7ae0b9dd7293bfa4157bdab" exitCode=0 Oct 10 09:52:49 crc kubenswrapper[4669]: I1010 09:52:49.402269 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" event={"ID":"5dc32d38-3a9b-48bb-93c4-76f51d215213","Type":"ContainerDied","Data":"a5d90cb8e32cc650179cace4a019a84dc29edc8bd7ae0b9dd7293bfa4157bdab"} Oct 10 09:52:49 crc kubenswrapper[4669]: I1010 09:52:49.796367 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:52:49 crc kubenswrapper[4669]: E1010 09:52:49.796953 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.804966 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.849497 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gch7\" (UniqueName: \"kubernetes.io/projected/5dc32d38-3a9b-48bb-93c4-76f51d215213-kube-api-access-2gch7\") pod \"5dc32d38-3a9b-48bb-93c4-76f51d215213\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.849575 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-inventory-0\") pod \"5dc32d38-3a9b-48bb-93c4-76f51d215213\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.849776 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ssh-key-openstack-edpm-ipam\") pod \"5dc32d38-3a9b-48bb-93c4-76f51d215213\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.849885 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ceph\") pod \"5dc32d38-3a9b-48bb-93c4-76f51d215213\" (UID: \"5dc32d38-3a9b-48bb-93c4-76f51d215213\") " Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.855489 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ceph" (OuterVolumeSpecName: "ceph") pod "5dc32d38-3a9b-48bb-93c4-76f51d215213" (UID: "5dc32d38-3a9b-48bb-93c4-76f51d215213"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.855693 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc32d38-3a9b-48bb-93c4-76f51d215213-kube-api-access-2gch7" (OuterVolumeSpecName: "kube-api-access-2gch7") pod "5dc32d38-3a9b-48bb-93c4-76f51d215213" (UID: "5dc32d38-3a9b-48bb-93c4-76f51d215213"). InnerVolumeSpecName "kube-api-access-2gch7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.881473 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5dc32d38-3a9b-48bb-93c4-76f51d215213" (UID: "5dc32d38-3a9b-48bb-93c4-76f51d215213"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.902552 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5dc32d38-3a9b-48bb-93c4-76f51d215213" (UID: "5dc32d38-3a9b-48bb-93c4-76f51d215213"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.953237 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.953268 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.953277 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gch7\" (UniqueName: \"kubernetes.io/projected/5dc32d38-3a9b-48bb-93c4-76f51d215213-kube-api-access-2gch7\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:50 crc kubenswrapper[4669]: I1010 09:52:50.953286 4669 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5dc32d38-3a9b-48bb-93c4-76f51d215213-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.440205 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" event={"ID":"5dc32d38-3a9b-48bb-93c4-76f51d215213","Type":"ContainerDied","Data":"47d49bbb62b4ab8d9ca9f63329b156c1b8582234616159f4dcf661e1ad3a9b28"} Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.440775 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47d49bbb62b4ab8d9ca9f63329b156c1b8582234616159f4dcf661e1ad3a9b28" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.440782 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dpnrk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.524613 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk"] Oct 10 09:52:51 crc kubenswrapper[4669]: E1010 09:52:51.525162 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc32d38-3a9b-48bb-93c4-76f51d215213" containerName="ssh-known-hosts-edpm-deployment" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.525225 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc32d38-3a9b-48bb-93c4-76f51d215213" containerName="ssh-known-hosts-edpm-deployment" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.525440 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc32d38-3a9b-48bb-93c4-76f51d215213" containerName="ssh-known-hosts-edpm-deployment" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.526057 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.527839 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.529896 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.530194 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.530200 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.537434 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.540723 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk"] Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.565975 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.566103 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwbfc\" (UniqueName: \"kubernetes.io/projected/483acfce-519d-4164-a1fe-e8786ade906f-kube-api-access-gwbfc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.566145 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.566248 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.667718 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwbfc\" (UniqueName: \"kubernetes.io/projected/483acfce-519d-4164-a1fe-e8786ade906f-kube-api-access-gwbfc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.667951 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.668073 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.668183 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.671408 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.672028 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.673979 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.682126 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwbfc\" (UniqueName: \"kubernetes.io/projected/483acfce-519d-4164-a1fe-e8786ade906f-kube-api-access-gwbfc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-768lk\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:51 crc kubenswrapper[4669]: I1010 09:52:51.894852 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:52:52 crc kubenswrapper[4669]: I1010 09:52:52.204310 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk"] Oct 10 09:52:52 crc kubenswrapper[4669]: W1010 09:52:52.212145 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod483acfce_519d_4164_a1fe_e8786ade906f.slice/crio-9d0cd29cb2715d4a72cffd3abb5a8c47e060254f5372a4f4daca4428be3d1e62 WatchSource:0}: Error finding container 9d0cd29cb2715d4a72cffd3abb5a8c47e060254f5372a4f4daca4428be3d1e62: Status 404 returned error can't find the container with id 9d0cd29cb2715d4a72cffd3abb5a8c47e060254f5372a4f4daca4428be3d1e62 Oct 10 09:52:52 crc kubenswrapper[4669]: I1010 09:52:52.450190 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" event={"ID":"483acfce-519d-4164-a1fe-e8786ade906f","Type":"ContainerStarted","Data":"9d0cd29cb2715d4a72cffd3abb5a8c47e060254f5372a4f4daca4428be3d1e62"} Oct 10 09:52:53 crc kubenswrapper[4669]: I1010 09:52:53.466677 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" event={"ID":"483acfce-519d-4164-a1fe-e8786ade906f","Type":"ContainerStarted","Data":"a9154f39016cea0adc20114bbd599caa9f5ee5800e2278e13383524953f3ac8d"} Oct 10 09:52:53 crc kubenswrapper[4669]: I1010 09:52:53.500628 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" podStartSLOduration=2.051670333 podStartE2EDuration="2.500607323s" podCreationTimestamp="2025-10-10 09:52:51 +0000 UTC" firstStartedPulling="2025-10-10 09:52:52.215168472 +0000 UTC m=+2515.231187214" lastFinishedPulling="2025-10-10 09:52:52.664105462 +0000 UTC m=+2515.680124204" observedRunningTime="2025-10-10 09:52:53.492661949 +0000 UTC m=+2516.508680731" watchObservedRunningTime="2025-10-10 09:52:53.500607323 +0000 UTC m=+2516.516626075" Oct 10 09:53:01 crc kubenswrapper[4669]: I1010 09:53:01.540365 4669 generic.go:334] "Generic (PLEG): container finished" podID="483acfce-519d-4164-a1fe-e8786ade906f" containerID="a9154f39016cea0adc20114bbd599caa9f5ee5800e2278e13383524953f3ac8d" exitCode=0 Oct 10 09:53:01 crc kubenswrapper[4669]: I1010 09:53:01.540445 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" event={"ID":"483acfce-519d-4164-a1fe-e8786ade906f","Type":"ContainerDied","Data":"a9154f39016cea0adc20114bbd599caa9f5ee5800e2278e13383524953f3ac8d"} Oct 10 09:53:01 crc kubenswrapper[4669]: I1010 09:53:01.796396 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:53:02 crc kubenswrapper[4669]: I1010 09:53:02.555858 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"17f3dc9423181b5b82c8eb0219ab63c353115b847743715862f372bb06c2885e"} Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.030759 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.115452 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwbfc\" (UniqueName: \"kubernetes.io/projected/483acfce-519d-4164-a1fe-e8786ade906f-kube-api-access-gwbfc\") pod \"483acfce-519d-4164-a1fe-e8786ade906f\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.115611 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ssh-key\") pod \"483acfce-519d-4164-a1fe-e8786ade906f\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.115648 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-inventory\") pod \"483acfce-519d-4164-a1fe-e8786ade906f\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.115750 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ceph\") pod \"483acfce-519d-4164-a1fe-e8786ade906f\" (UID: \"483acfce-519d-4164-a1fe-e8786ade906f\") " Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.120994 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ceph" (OuterVolumeSpecName: "ceph") pod "483acfce-519d-4164-a1fe-e8786ade906f" (UID: "483acfce-519d-4164-a1fe-e8786ade906f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.125709 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483acfce-519d-4164-a1fe-e8786ade906f-kube-api-access-gwbfc" (OuterVolumeSpecName: "kube-api-access-gwbfc") pod "483acfce-519d-4164-a1fe-e8786ade906f" (UID: "483acfce-519d-4164-a1fe-e8786ade906f"). InnerVolumeSpecName "kube-api-access-gwbfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.145239 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "483acfce-519d-4164-a1fe-e8786ade906f" (UID: "483acfce-519d-4164-a1fe-e8786ade906f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.147296 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-inventory" (OuterVolumeSpecName: "inventory") pod "483acfce-519d-4164-a1fe-e8786ade906f" (UID: "483acfce-519d-4164-a1fe-e8786ade906f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.218959 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.219009 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwbfc\" (UniqueName: \"kubernetes.io/projected/483acfce-519d-4164-a1fe-e8786ade906f-kube-api-access-gwbfc\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.219035 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.219053 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/483acfce-519d-4164-a1fe-e8786ade906f-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.565856 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" event={"ID":"483acfce-519d-4164-a1fe-e8786ade906f","Type":"ContainerDied","Data":"9d0cd29cb2715d4a72cffd3abb5a8c47e060254f5372a4f4daca4428be3d1e62"} Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.567081 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d0cd29cb2715d4a72cffd3abb5a8c47e060254f5372a4f4daca4428be3d1e62" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.565926 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-768lk" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.658015 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c"] Oct 10 09:53:03 crc kubenswrapper[4669]: E1010 09:53:03.658386 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483acfce-519d-4164-a1fe-e8786ade906f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.658405 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="483acfce-519d-4164-a1fe-e8786ade906f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.658673 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="483acfce-519d-4164-a1fe-e8786ade906f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.659353 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.663105 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.663425 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.670847 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.670945 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.671194 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.682798 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c"] Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.728069 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.728124 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw5ml\" (UniqueName: \"kubernetes.io/projected/f9938853-891c-4766-82ef-21e7b718dcc2-kube-api-access-zw5ml\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.728415 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.728529 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.829939 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.830016 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.830079 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.830107 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw5ml\" (UniqueName: \"kubernetes.io/projected/f9938853-891c-4766-82ef-21e7b718dcc2-kube-api-access-zw5ml\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.835320 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.835578 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.839682 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.853347 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw5ml\" (UniqueName: \"kubernetes.io/projected/f9938853-891c-4766-82ef-21e7b718dcc2-kube-api-access-zw5ml\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:03 crc kubenswrapper[4669]: I1010 09:53:03.987720 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:04 crc kubenswrapper[4669]: I1010 09:53:04.562900 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c"] Oct 10 09:53:04 crc kubenswrapper[4669]: W1010 09:53:04.574759 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9938853_891c_4766_82ef_21e7b718dcc2.slice/crio-c4dc5e8a1623a0ba28ac8eb7183fe88b40e95370d61696ea7d2227020c55770f WatchSource:0}: Error finding container c4dc5e8a1623a0ba28ac8eb7183fe88b40e95370d61696ea7d2227020c55770f: Status 404 returned error can't find the container with id c4dc5e8a1623a0ba28ac8eb7183fe88b40e95370d61696ea7d2227020c55770f Oct 10 09:53:05 crc kubenswrapper[4669]: I1010 09:53:05.595831 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" event={"ID":"f9938853-891c-4766-82ef-21e7b718dcc2","Type":"ContainerStarted","Data":"5ee29b0dee8fc6005ce98019ee36acef0646d3361edff2f1164362b5d36e57e7"} Oct 10 09:53:05 crc kubenswrapper[4669]: I1010 09:53:05.596313 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" event={"ID":"f9938853-891c-4766-82ef-21e7b718dcc2","Type":"ContainerStarted","Data":"c4dc5e8a1623a0ba28ac8eb7183fe88b40e95370d61696ea7d2227020c55770f"} Oct 10 09:53:05 crc kubenswrapper[4669]: I1010 09:53:05.621553 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" podStartSLOduration=1.951123997 podStartE2EDuration="2.621533032s" podCreationTimestamp="2025-10-10 09:53:03 +0000 UTC" firstStartedPulling="2025-10-10 09:53:04.579429595 +0000 UTC m=+2527.595448337" lastFinishedPulling="2025-10-10 09:53:05.24983859 +0000 UTC m=+2528.265857372" observedRunningTime="2025-10-10 09:53:05.616877393 +0000 UTC m=+2528.632896145" watchObservedRunningTime="2025-10-10 09:53:05.621533032 +0000 UTC m=+2528.637551794" Oct 10 09:53:15 crc kubenswrapper[4669]: I1010 09:53:15.687665 4669 generic.go:334] "Generic (PLEG): container finished" podID="f9938853-891c-4766-82ef-21e7b718dcc2" containerID="5ee29b0dee8fc6005ce98019ee36acef0646d3361edff2f1164362b5d36e57e7" exitCode=0 Oct 10 09:53:15 crc kubenswrapper[4669]: I1010 09:53:15.687726 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" event={"ID":"f9938853-891c-4766-82ef-21e7b718dcc2","Type":"ContainerDied","Data":"5ee29b0dee8fc6005ce98019ee36acef0646d3361edff2f1164362b5d36e57e7"} Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.142294 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.210287 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw5ml\" (UniqueName: \"kubernetes.io/projected/f9938853-891c-4766-82ef-21e7b718dcc2-kube-api-access-zw5ml\") pod \"f9938853-891c-4766-82ef-21e7b718dcc2\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.210543 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-inventory\") pod \"f9938853-891c-4766-82ef-21e7b718dcc2\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.210615 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ssh-key\") pod \"f9938853-891c-4766-82ef-21e7b718dcc2\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.210676 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ceph\") pod \"f9938853-891c-4766-82ef-21e7b718dcc2\" (UID: \"f9938853-891c-4766-82ef-21e7b718dcc2\") " Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.219495 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ceph" (OuterVolumeSpecName: "ceph") pod "f9938853-891c-4766-82ef-21e7b718dcc2" (UID: "f9938853-891c-4766-82ef-21e7b718dcc2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.219696 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9938853-891c-4766-82ef-21e7b718dcc2-kube-api-access-zw5ml" (OuterVolumeSpecName: "kube-api-access-zw5ml") pod "f9938853-891c-4766-82ef-21e7b718dcc2" (UID: "f9938853-891c-4766-82ef-21e7b718dcc2"). InnerVolumeSpecName "kube-api-access-zw5ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.236975 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9938853-891c-4766-82ef-21e7b718dcc2" (UID: "f9938853-891c-4766-82ef-21e7b718dcc2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.241158 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-inventory" (OuterVolumeSpecName: "inventory") pod "f9938853-891c-4766-82ef-21e7b718dcc2" (UID: "f9938853-891c-4766-82ef-21e7b718dcc2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.313069 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.313134 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.313158 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9938853-891c-4766-82ef-21e7b718dcc2-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.313180 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw5ml\" (UniqueName: \"kubernetes.io/projected/f9938853-891c-4766-82ef-21e7b718dcc2-kube-api-access-zw5ml\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.714872 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" event={"ID":"f9938853-891c-4766-82ef-21e7b718dcc2","Type":"ContainerDied","Data":"c4dc5e8a1623a0ba28ac8eb7183fe88b40e95370d61696ea7d2227020c55770f"} Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.714918 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4dc5e8a1623a0ba28ac8eb7183fe88b40e95370d61696ea7d2227020c55770f" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.715551 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.848747 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b"] Oct 10 09:53:17 crc kubenswrapper[4669]: E1010 09:53:17.849108 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9938853-891c-4766-82ef-21e7b718dcc2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.849126 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9938853-891c-4766-82ef-21e7b718dcc2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.849312 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9938853-891c-4766-82ef-21e7b718dcc2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.850042 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.853170 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.853191 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.853310 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.853383 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.853478 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.853575 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.853724 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.855923 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.906435 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b"] Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928357 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928400 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928431 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928498 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928612 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928717 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928766 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928792 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928839 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928878 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7djn7\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-kube-api-access-7djn7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928914 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.928957 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:17 crc kubenswrapper[4669]: I1010 09:53:17.929013 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.030837 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031141 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031192 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031217 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031235 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031256 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031282 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031302 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031330 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031351 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031374 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031408 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.031434 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7djn7\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-kube-api-access-7djn7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.036294 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.036783 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.037613 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.038050 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.038323 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.037627 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.039727 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.040266 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.041898 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.042793 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.042880 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.044184 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.055376 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7djn7\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-kube-api-access-7djn7\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.171092 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.693808 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b"] Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.700659 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 09:53:18 crc kubenswrapper[4669]: I1010 09:53:18.722167 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" event={"ID":"880204d3-a81a-433f-ace4-ec7ff0092b53","Type":"ContainerStarted","Data":"fbf9d1270e6cbce1faa8677af01358c5cd995aa6b55d2030fbe826e039960bb6"} Oct 10 09:53:19 crc kubenswrapper[4669]: I1010 09:53:19.733239 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" event={"ID":"880204d3-a81a-433f-ace4-ec7ff0092b53","Type":"ContainerStarted","Data":"59ed1d634b784fa38437da71a2be56f34cd19482a87ec0b6091af6e54042ee35"} Oct 10 09:53:19 crc kubenswrapper[4669]: I1010 09:53:19.756417 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" podStartSLOduration=2.257616902 podStartE2EDuration="2.756398831s" podCreationTimestamp="2025-10-10 09:53:17 +0000 UTC" firstStartedPulling="2025-10-10 09:53:18.7004426 +0000 UTC m=+2541.716461342" lastFinishedPulling="2025-10-10 09:53:19.199224519 +0000 UTC m=+2542.215243271" observedRunningTime="2025-10-10 09:53:19.749468059 +0000 UTC m=+2542.765486811" watchObservedRunningTime="2025-10-10 09:53:19.756398831 +0000 UTC m=+2542.772417573" Oct 10 09:53:57 crc kubenswrapper[4669]: I1010 09:53:57.055722 4669 generic.go:334] "Generic (PLEG): container finished" podID="880204d3-a81a-433f-ace4-ec7ff0092b53" containerID="59ed1d634b784fa38437da71a2be56f34cd19482a87ec0b6091af6e54042ee35" exitCode=0 Oct 10 09:53:57 crc kubenswrapper[4669]: I1010 09:53:57.055900 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" event={"ID":"880204d3-a81a-433f-ace4-ec7ff0092b53","Type":"ContainerDied","Data":"59ed1d634b784fa38437da71a2be56f34cd19482a87ec0b6091af6e54042ee35"} Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.714412 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800524 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-libvirt-combined-ca-bundle\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800574 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-repo-setup-combined-ca-bundle\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800634 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-nova-combined-ca-bundle\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800691 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ovn-combined-ca-bundle\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800710 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7djn7\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-kube-api-access-7djn7\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800775 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800802 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-bootstrap-combined-ca-bundle\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800899 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.800936 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ssh-key\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.801018 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-neutron-metadata-combined-ca-bundle\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.801041 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-inventory\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.801082 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ceph\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.801122 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-ovn-default-certs-0\") pod \"880204d3-a81a-433f-ace4-ec7ff0092b53\" (UID: \"880204d3-a81a-433f-ace4-ec7ff0092b53\") " Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.806900 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.810108 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.810894 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.811344 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.811332 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.812877 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ceph" (OuterVolumeSpecName: "ceph") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.813628 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.814014 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.814244 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-kube-api-access-7djn7" (OuterVolumeSpecName: "kube-api-access-7djn7") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "kube-api-access-7djn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.814620 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.819360 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.841445 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-inventory" (OuterVolumeSpecName: "inventory") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.848035 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "880204d3-a81a-433f-ace4-ec7ff0092b53" (UID: "880204d3-a81a-433f-ace4-ec7ff0092b53"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905761 4669 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905799 4669 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905809 4669 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905821 4669 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905829 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7djn7\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-kube-api-access-7djn7\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905844 4669 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905855 4669 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905864 4669 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905874 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905882 4669 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905912 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905938 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/880204d3-a81a-433f-ace4-ec7ff0092b53-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:58 crc kubenswrapper[4669]: I1010 09:53:58.905948 4669 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/880204d3-a81a-433f-ace4-ec7ff0092b53-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.076126 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" event={"ID":"880204d3-a81a-433f-ace4-ec7ff0092b53","Type":"ContainerDied","Data":"fbf9d1270e6cbce1faa8677af01358c5cd995aa6b55d2030fbe826e039960bb6"} Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.076168 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbf9d1270e6cbce1faa8677af01358c5cd995aa6b55d2030fbe826e039960bb6" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.076227 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.182790 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q"] Oct 10 09:53:59 crc kubenswrapper[4669]: E1010 09:53:59.183454 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880204d3-a81a-433f-ace4-ec7ff0092b53" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.183478 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="880204d3-a81a-433f-ace4-ec7ff0092b53" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.183747 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="880204d3-a81a-433f-ace4-ec7ff0092b53" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.184495 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.187430 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.187469 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.187656 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.195632 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.195847 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.195953 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q"] Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.311792 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.311859 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wwcg\" (UniqueName: \"kubernetes.io/projected/c1b45643-fd51-4a37-a298-f9b84b997c40-kube-api-access-8wwcg\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.311998 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.312074 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.414023 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wwcg\" (UniqueName: \"kubernetes.io/projected/c1b45643-fd51-4a37-a298-f9b84b997c40-kube-api-access-8wwcg\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.414091 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.414157 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.414217 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.418435 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.420316 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.425161 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.435172 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wwcg\" (UniqueName: \"kubernetes.io/projected/c1b45643-fd51-4a37-a298-f9b84b997c40-kube-api-access-8wwcg\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:53:59 crc kubenswrapper[4669]: I1010 09:53:59.517643 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:54:00 crc kubenswrapper[4669]: I1010 09:54:00.065454 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q"] Oct 10 09:54:00 crc kubenswrapper[4669]: I1010 09:54:00.085435 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" event={"ID":"c1b45643-fd51-4a37-a298-f9b84b997c40","Type":"ContainerStarted","Data":"9684de31d94764b22da0c650c5d244034614fded60726c02971b3d7b289df1b3"} Oct 10 09:54:01 crc kubenswrapper[4669]: I1010 09:54:01.094812 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" event={"ID":"c1b45643-fd51-4a37-a298-f9b84b997c40","Type":"ContainerStarted","Data":"2eadce3e589b5c05a081d0dd25edb29cf7c229274b5ccf191b379adce1a9ea70"} Oct 10 09:54:01 crc kubenswrapper[4669]: I1010 09:54:01.121689 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" podStartSLOduration=1.6230648140000001 podStartE2EDuration="2.121672207s" podCreationTimestamp="2025-10-10 09:53:59 +0000 UTC" firstStartedPulling="2025-10-10 09:54:00.068458654 +0000 UTC m=+2583.084477396" lastFinishedPulling="2025-10-10 09:54:00.567066007 +0000 UTC m=+2583.583084789" observedRunningTime="2025-10-10 09:54:01.114754944 +0000 UTC m=+2584.130773686" watchObservedRunningTime="2025-10-10 09:54:01.121672207 +0000 UTC m=+2584.137690949" Oct 10 09:54:07 crc kubenswrapper[4669]: I1010 09:54:07.150497 4669 generic.go:334] "Generic (PLEG): container finished" podID="c1b45643-fd51-4a37-a298-f9b84b997c40" containerID="2eadce3e589b5c05a081d0dd25edb29cf7c229274b5ccf191b379adce1a9ea70" exitCode=0 Oct 10 09:54:07 crc kubenswrapper[4669]: I1010 09:54:07.150629 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" event={"ID":"c1b45643-fd51-4a37-a298-f9b84b997c40","Type":"ContainerDied","Data":"2eadce3e589b5c05a081d0dd25edb29cf7c229274b5ccf191b379adce1a9ea70"} Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.606510 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.687052 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-inventory\") pod \"c1b45643-fd51-4a37-a298-f9b84b997c40\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.689071 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ceph\") pod \"c1b45643-fd51-4a37-a298-f9b84b997c40\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.689264 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wwcg\" (UniqueName: \"kubernetes.io/projected/c1b45643-fd51-4a37-a298-f9b84b997c40-kube-api-access-8wwcg\") pod \"c1b45643-fd51-4a37-a298-f9b84b997c40\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.689662 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ssh-key\") pod \"c1b45643-fd51-4a37-a298-f9b84b997c40\" (UID: \"c1b45643-fd51-4a37-a298-f9b84b997c40\") " Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.694299 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b45643-fd51-4a37-a298-f9b84b997c40-kube-api-access-8wwcg" (OuterVolumeSpecName: "kube-api-access-8wwcg") pod "c1b45643-fd51-4a37-a298-f9b84b997c40" (UID: "c1b45643-fd51-4a37-a298-f9b84b997c40"). InnerVolumeSpecName "kube-api-access-8wwcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.709102 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ceph" (OuterVolumeSpecName: "ceph") pod "c1b45643-fd51-4a37-a298-f9b84b997c40" (UID: "c1b45643-fd51-4a37-a298-f9b84b997c40"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.716759 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-inventory" (OuterVolumeSpecName: "inventory") pod "c1b45643-fd51-4a37-a298-f9b84b997c40" (UID: "c1b45643-fd51-4a37-a298-f9b84b997c40"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.717368 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c1b45643-fd51-4a37-a298-f9b84b997c40" (UID: "c1b45643-fd51-4a37-a298-f9b84b997c40"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.791923 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.791966 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.791978 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c1b45643-fd51-4a37-a298-f9b84b997c40-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:54:08 crc kubenswrapper[4669]: I1010 09:54:08.791988 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wwcg\" (UniqueName: \"kubernetes.io/projected/c1b45643-fd51-4a37-a298-f9b84b997c40-kube-api-access-8wwcg\") on node \"crc\" DevicePath \"\"" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.189295 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.189455 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q" event={"ID":"c1b45643-fd51-4a37-a298-f9b84b997c40","Type":"ContainerDied","Data":"9684de31d94764b22da0c650c5d244034614fded60726c02971b3d7b289df1b3"} Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.189485 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9684de31d94764b22da0c650c5d244034614fded60726c02971b3d7b289df1b3" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.256127 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k"] Oct 10 09:54:09 crc kubenswrapper[4669]: E1010 09:54:09.256469 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b45643-fd51-4a37-a298-f9b84b997c40" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.256484 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b45643-fd51-4a37-a298-f9b84b997c40" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.256717 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b45643-fd51-4a37-a298-f9b84b997c40" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.257320 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.259744 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.260635 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.261188 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.261329 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.262926 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.265640 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.278233 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k"] Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.300958 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.301017 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.301049 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cd3df673-494f-4e69-a140-34e78f12567c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.301082 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.301098 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drwng\" (UniqueName: \"kubernetes.io/projected/cd3df673-494f-4e69-a140-34e78f12567c-kube-api-access-drwng\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.301121 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.402532 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.402628 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.402663 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cd3df673-494f-4e69-a140-34e78f12567c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.402716 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.402738 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drwng\" (UniqueName: \"kubernetes.io/projected/cd3df673-494f-4e69-a140-34e78f12567c-kube-api-access-drwng\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.402784 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.403538 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cd3df673-494f-4e69-a140-34e78f12567c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.407390 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.407446 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.407463 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.407529 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.423632 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drwng\" (UniqueName: \"kubernetes.io/projected/cd3df673-494f-4e69-a140-34e78f12567c-kube-api-access-drwng\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-nsc5k\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:09 crc kubenswrapper[4669]: I1010 09:54:09.575516 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:54:10 crc kubenswrapper[4669]: I1010 09:54:10.079558 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k"] Oct 10 09:54:10 crc kubenswrapper[4669]: W1010 09:54:10.080598 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd3df673_494f_4e69_a140_34e78f12567c.slice/crio-3d4b7a80cb58fcbf4fe1af499647e92b0ef09cd33abf85bcda3bd868c0a43cde WatchSource:0}: Error finding container 3d4b7a80cb58fcbf4fe1af499647e92b0ef09cd33abf85bcda3bd868c0a43cde: Status 404 returned error can't find the container with id 3d4b7a80cb58fcbf4fe1af499647e92b0ef09cd33abf85bcda3bd868c0a43cde Oct 10 09:54:10 crc kubenswrapper[4669]: I1010 09:54:10.197169 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" event={"ID":"cd3df673-494f-4e69-a140-34e78f12567c","Type":"ContainerStarted","Data":"3d4b7a80cb58fcbf4fe1af499647e92b0ef09cd33abf85bcda3bd868c0a43cde"} Oct 10 09:54:11 crc kubenswrapper[4669]: I1010 09:54:11.210118 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" event={"ID":"cd3df673-494f-4e69-a140-34e78f12567c","Type":"ContainerStarted","Data":"4cf75f726b9e1b82d3b378bfbe1add8d046b76e4c076ab84dca38b2436304098"} Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.213379 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" podStartSLOduration=54.719154654 podStartE2EDuration="55.213351725s" podCreationTimestamp="2025-10-10 09:54:09 +0000 UTC" firstStartedPulling="2025-10-10 09:54:10.083155344 +0000 UTC m=+2593.099174086" lastFinishedPulling="2025-10-10 09:54:10.577352375 +0000 UTC m=+2593.593371157" observedRunningTime="2025-10-10 09:54:11.239398531 +0000 UTC m=+2594.255417273" watchObservedRunningTime="2025-10-10 09:55:04.213351725 +0000 UTC m=+2647.229370467" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.218100 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sg5rb"] Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.220743 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.232563 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sg5rb"] Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.277752 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xqkc\" (UniqueName: \"kubernetes.io/projected/16edfa98-b699-403b-ae5e-30958ece9e1e-kube-api-access-9xqkc\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.277831 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-utilities\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.277867 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-catalog-content\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.379766 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-utilities\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.379847 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-catalog-content\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.380324 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-utilities\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.380355 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-catalog-content\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.380869 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xqkc\" (UniqueName: \"kubernetes.io/projected/16edfa98-b699-403b-ae5e-30958ece9e1e-kube-api-access-9xqkc\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.399098 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xqkc\" (UniqueName: \"kubernetes.io/projected/16edfa98-b699-403b-ae5e-30958ece9e1e-kube-api-access-9xqkc\") pod \"community-operators-sg5rb\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:04 crc kubenswrapper[4669]: I1010 09:55:04.589485 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:05 crc kubenswrapper[4669]: I1010 09:55:05.177291 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sg5rb"] Oct 10 09:55:05 crc kubenswrapper[4669]: W1010 09:55:05.180140 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16edfa98_b699_403b_ae5e_30958ece9e1e.slice/crio-b0f05514085589723614055abfe8765ace1ca9d4e87f9473c61378a9ee643ecd WatchSource:0}: Error finding container b0f05514085589723614055abfe8765ace1ca9d4e87f9473c61378a9ee643ecd: Status 404 returned error can't find the container with id b0f05514085589723614055abfe8765ace1ca9d4e87f9473c61378a9ee643ecd Oct 10 09:55:05 crc kubenswrapper[4669]: I1010 09:55:05.737958 4669 generic.go:334] "Generic (PLEG): container finished" podID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerID="050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6" exitCode=0 Oct 10 09:55:05 crc kubenswrapper[4669]: I1010 09:55:05.738492 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5rb" event={"ID":"16edfa98-b699-403b-ae5e-30958ece9e1e","Type":"ContainerDied","Data":"050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6"} Oct 10 09:55:05 crc kubenswrapper[4669]: I1010 09:55:05.738522 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5rb" event={"ID":"16edfa98-b699-403b-ae5e-30958ece9e1e","Type":"ContainerStarted","Data":"b0f05514085589723614055abfe8765ace1ca9d4e87f9473c61378a9ee643ecd"} Oct 10 09:55:07 crc kubenswrapper[4669]: I1010 09:55:07.771773 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5rb" event={"ID":"16edfa98-b699-403b-ae5e-30958ece9e1e","Type":"ContainerStarted","Data":"02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0"} Oct 10 09:55:08 crc kubenswrapper[4669]: I1010 09:55:08.785364 4669 generic.go:334] "Generic (PLEG): container finished" podID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerID="02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0" exitCode=0 Oct 10 09:55:08 crc kubenswrapper[4669]: I1010 09:55:08.785447 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5rb" event={"ID":"16edfa98-b699-403b-ae5e-30958ece9e1e","Type":"ContainerDied","Data":"02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0"} Oct 10 09:55:09 crc kubenswrapper[4669]: I1010 09:55:09.803527 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5rb" event={"ID":"16edfa98-b699-403b-ae5e-30958ece9e1e","Type":"ContainerStarted","Data":"befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552"} Oct 10 09:55:09 crc kubenswrapper[4669]: I1010 09:55:09.818764 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sg5rb" podStartSLOduration=2.315623338 podStartE2EDuration="5.818738863s" podCreationTimestamp="2025-10-10 09:55:04 +0000 UTC" firstStartedPulling="2025-10-10 09:55:05.741697319 +0000 UTC m=+2648.757716061" lastFinishedPulling="2025-10-10 09:55:09.244812844 +0000 UTC m=+2652.260831586" observedRunningTime="2025-10-10 09:55:09.817192524 +0000 UTC m=+2652.833211266" watchObservedRunningTime="2025-10-10 09:55:09.818738863 +0000 UTC m=+2652.834757605" Oct 10 09:55:14 crc kubenswrapper[4669]: I1010 09:55:14.590414 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:14 crc kubenswrapper[4669]: I1010 09:55:14.591005 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:14 crc kubenswrapper[4669]: I1010 09:55:14.651258 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:14 crc kubenswrapper[4669]: I1010 09:55:14.897017 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:14 crc kubenswrapper[4669]: I1010 09:55:14.946026 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sg5rb"] Oct 10 09:55:16 crc kubenswrapper[4669]: I1010 09:55:16.863793 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sg5rb" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="registry-server" containerID="cri-o://befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552" gracePeriod=2 Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.357872 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.534325 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-utilities\") pod \"16edfa98-b699-403b-ae5e-30958ece9e1e\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.534509 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xqkc\" (UniqueName: \"kubernetes.io/projected/16edfa98-b699-403b-ae5e-30958ece9e1e-kube-api-access-9xqkc\") pod \"16edfa98-b699-403b-ae5e-30958ece9e1e\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.534572 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-catalog-content\") pod \"16edfa98-b699-403b-ae5e-30958ece9e1e\" (UID: \"16edfa98-b699-403b-ae5e-30958ece9e1e\") " Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.535713 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-utilities" (OuterVolumeSpecName: "utilities") pod "16edfa98-b699-403b-ae5e-30958ece9e1e" (UID: "16edfa98-b699-403b-ae5e-30958ece9e1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.540812 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16edfa98-b699-403b-ae5e-30958ece9e1e-kube-api-access-9xqkc" (OuterVolumeSpecName: "kube-api-access-9xqkc") pod "16edfa98-b699-403b-ae5e-30958ece9e1e" (UID: "16edfa98-b699-403b-ae5e-30958ece9e1e"). InnerVolumeSpecName "kube-api-access-9xqkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.578006 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16edfa98-b699-403b-ae5e-30958ece9e1e" (UID: "16edfa98-b699-403b-ae5e-30958ece9e1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.637555 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.637883 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xqkc\" (UniqueName: \"kubernetes.io/projected/16edfa98-b699-403b-ae5e-30958ece9e1e-kube-api-access-9xqkc\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.637954 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16edfa98-b699-403b-ae5e-30958ece9e1e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.873179 4669 generic.go:334] "Generic (PLEG): container finished" podID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerID="befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552" exitCode=0 Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.873219 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5rb" event={"ID":"16edfa98-b699-403b-ae5e-30958ece9e1e","Type":"ContainerDied","Data":"befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552"} Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.873244 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg5rb" event={"ID":"16edfa98-b699-403b-ae5e-30958ece9e1e","Type":"ContainerDied","Data":"b0f05514085589723614055abfe8765ace1ca9d4e87f9473c61378a9ee643ecd"} Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.873260 4669 scope.go:117] "RemoveContainer" containerID="befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.873402 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg5rb" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.909462 4669 scope.go:117] "RemoveContainer" containerID="02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.909962 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sg5rb"] Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.918430 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sg5rb"] Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.940652 4669 scope.go:117] "RemoveContainer" containerID="050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.969082 4669 scope.go:117] "RemoveContainer" containerID="befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552" Oct 10 09:55:17 crc kubenswrapper[4669]: E1010 09:55:17.969688 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552\": container with ID starting with befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552 not found: ID does not exist" containerID="befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.969729 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552"} err="failed to get container status \"befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552\": rpc error: code = NotFound desc = could not find container \"befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552\": container with ID starting with befe1d5fcea217c9134abf8d8f59fca5772b04bd20bc178c18a357f136b16552 not found: ID does not exist" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.969759 4669 scope.go:117] "RemoveContainer" containerID="02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0" Oct 10 09:55:17 crc kubenswrapper[4669]: E1010 09:55:17.970231 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0\": container with ID starting with 02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0 not found: ID does not exist" containerID="02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.970264 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0"} err="failed to get container status \"02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0\": rpc error: code = NotFound desc = could not find container \"02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0\": container with ID starting with 02821f5ff104da9d788aba01be59ec3739df7f0a86bcfe1501d8e36f1a4549e0 not found: ID does not exist" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.970282 4669 scope.go:117] "RemoveContainer" containerID="050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6" Oct 10 09:55:17 crc kubenswrapper[4669]: E1010 09:55:17.970532 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6\": container with ID starting with 050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6 not found: ID does not exist" containerID="050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6" Oct 10 09:55:17 crc kubenswrapper[4669]: I1010 09:55:17.970563 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6"} err="failed to get container status \"050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6\": rpc error: code = NotFound desc = could not find container \"050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6\": container with ID starting with 050c98dc918bce62904644f7589f33d398a2fd4c4fa40a5345790b82014c78e6 not found: ID does not exist" Oct 10 09:55:19 crc kubenswrapper[4669]: I1010 09:55:19.805359 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" path="/var/lib/kubelet/pods/16edfa98-b699-403b-ae5e-30958ece9e1e/volumes" Oct 10 09:55:24 crc kubenswrapper[4669]: I1010 09:55:24.275173 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:55:24 crc kubenswrapper[4669]: I1010 09:55:24.276728 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:55:31 crc kubenswrapper[4669]: I1010 09:55:31.994944 4669 generic.go:334] "Generic (PLEG): container finished" podID="cd3df673-494f-4e69-a140-34e78f12567c" containerID="4cf75f726b9e1b82d3b378bfbe1add8d046b76e4c076ab84dca38b2436304098" exitCode=0 Oct 10 09:55:31 crc kubenswrapper[4669]: I1010 09:55:31.995138 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" event={"ID":"cd3df673-494f-4e69-a140-34e78f12567c","Type":"ContainerDied","Data":"4cf75f726b9e1b82d3b378bfbe1add8d046b76e4c076ab84dca38b2436304098"} Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.411038 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.581494 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ssh-key\") pod \"cd3df673-494f-4e69-a140-34e78f12567c\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.581651 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-inventory\") pod \"cd3df673-494f-4e69-a140-34e78f12567c\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.581829 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ovn-combined-ca-bundle\") pod \"cd3df673-494f-4e69-a140-34e78f12567c\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.582364 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drwng\" (UniqueName: \"kubernetes.io/projected/cd3df673-494f-4e69-a140-34e78f12567c-kube-api-access-drwng\") pod \"cd3df673-494f-4e69-a140-34e78f12567c\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.582407 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cd3df673-494f-4e69-a140-34e78f12567c-ovncontroller-config-0\") pod \"cd3df673-494f-4e69-a140-34e78f12567c\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.582513 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ceph\") pod \"cd3df673-494f-4e69-a140-34e78f12567c\" (UID: \"cd3df673-494f-4e69-a140-34e78f12567c\") " Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.587575 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd3df673-494f-4e69-a140-34e78f12567c-kube-api-access-drwng" (OuterVolumeSpecName: "kube-api-access-drwng") pod "cd3df673-494f-4e69-a140-34e78f12567c" (UID: "cd3df673-494f-4e69-a140-34e78f12567c"). InnerVolumeSpecName "kube-api-access-drwng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.588974 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ceph" (OuterVolumeSpecName: "ceph") pod "cd3df673-494f-4e69-a140-34e78f12567c" (UID: "cd3df673-494f-4e69-a140-34e78f12567c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.590929 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "cd3df673-494f-4e69-a140-34e78f12567c" (UID: "cd3df673-494f-4e69-a140-34e78f12567c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.612527 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cd3df673-494f-4e69-a140-34e78f12567c" (UID: "cd3df673-494f-4e69-a140-34e78f12567c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.613101 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-inventory" (OuterVolumeSpecName: "inventory") pod "cd3df673-494f-4e69-a140-34e78f12567c" (UID: "cd3df673-494f-4e69-a140-34e78f12567c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.627804 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd3df673-494f-4e69-a140-34e78f12567c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "cd3df673-494f-4e69-a140-34e78f12567c" (UID: "cd3df673-494f-4e69-a140-34e78f12567c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.684802 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.684834 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.684845 4669 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.684856 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drwng\" (UniqueName: \"kubernetes.io/projected/cd3df673-494f-4e69-a140-34e78f12567c-kube-api-access-drwng\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.684867 4669 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/cd3df673-494f-4e69-a140-34e78f12567c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:33 crc kubenswrapper[4669]: I1010 09:55:33.684875 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cd3df673-494f-4e69-a140-34e78f12567c-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.018533 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" event={"ID":"cd3df673-494f-4e69-a140-34e78f12567c","Type":"ContainerDied","Data":"3d4b7a80cb58fcbf4fe1af499647e92b0ef09cd33abf85bcda3bd868c0a43cde"} Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.018605 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d4b7a80cb58fcbf4fe1af499647e92b0ef09cd33abf85bcda3bd868c0a43cde" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.019036 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-nsc5k" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.101886 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w"] Oct 10 09:55:34 crc kubenswrapper[4669]: E1010 09:55:34.102866 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd3df673-494f-4e69-a140-34e78f12567c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.103012 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd3df673-494f-4e69-a140-34e78f12567c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 10 09:55:34 crc kubenswrapper[4669]: E1010 09:55:34.103154 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="extract-content" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.103300 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="extract-content" Oct 10 09:55:34 crc kubenswrapper[4669]: E1010 09:55:34.103449 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="registry-server" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.103576 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="registry-server" Oct 10 09:55:34 crc kubenswrapper[4669]: E1010 09:55:34.103793 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="extract-utilities" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.103974 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="extract-utilities" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.104433 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd3df673-494f-4e69-a140-34e78f12567c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.104632 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="16edfa98-b699-403b-ae5e-30958ece9e1e" containerName="registry-server" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.105936 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.110483 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.110773 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.110947 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.111099 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.111240 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w"] Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.111297 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.114192 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.114192 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.295417 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.295476 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.295511 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.295555 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.295584 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.295640 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8tpz\" (UniqueName: \"kubernetes.io/projected/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-kube-api-access-n8tpz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.295673 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.398757 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.398810 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.398842 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.399401 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.399444 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.399491 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8tpz\" (UniqueName: \"kubernetes.io/projected/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-kube-api-access-n8tpz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.399549 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.405664 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.405728 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.405813 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.406985 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.409904 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.419257 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.424926 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8tpz\" (UniqueName: \"kubernetes.io/projected/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-kube-api-access-n8tpz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.425972 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:55:34 crc kubenswrapper[4669]: I1010 09:55:34.953632 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w"] Oct 10 09:55:35 crc kubenswrapper[4669]: I1010 09:55:35.026518 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" event={"ID":"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e","Type":"ContainerStarted","Data":"e4e7e132f3e4d6eb012a9f69c47e402e3da2d432d20abe217b6d73b67fb17192"} Oct 10 09:55:36 crc kubenswrapper[4669]: I1010 09:55:36.039151 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" event={"ID":"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e","Type":"ContainerStarted","Data":"c853cf72a7b476d87a0be0c8ee3bc29935cd6aa32fbd254442187226df57ed5e"} Oct 10 09:55:36 crc kubenswrapper[4669]: I1010 09:55:36.072208 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" podStartSLOduration=1.691763272 podStartE2EDuration="2.072147673s" podCreationTimestamp="2025-10-10 09:55:34 +0000 UTC" firstStartedPulling="2025-10-10 09:55:34.971572321 +0000 UTC m=+2677.987591063" lastFinishedPulling="2025-10-10 09:55:35.351956722 +0000 UTC m=+2678.367975464" observedRunningTime="2025-10-10 09:55:36.063188465 +0000 UTC m=+2679.079207207" watchObservedRunningTime="2025-10-10 09:55:36.072147673 +0000 UTC m=+2679.088166445" Oct 10 09:55:54 crc kubenswrapper[4669]: I1010 09:55:54.274784 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:55:54 crc kubenswrapper[4669]: I1010 09:55:54.275475 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.275448 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.276063 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.276118 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.276836 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17f3dc9423181b5b82c8eb0219ab63c353115b847743715862f372bb06c2885e"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.276881 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://17f3dc9423181b5b82c8eb0219ab63c353115b847743715862f372bb06c2885e" gracePeriod=600 Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.464419 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="17f3dc9423181b5b82c8eb0219ab63c353115b847743715862f372bb06c2885e" exitCode=0 Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.464916 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"17f3dc9423181b5b82c8eb0219ab63c353115b847743715862f372bb06c2885e"} Oct 10 09:56:24 crc kubenswrapper[4669]: I1010 09:56:24.465056 4669 scope.go:117] "RemoveContainer" containerID="f101c56873953c808e2aca493324baba04524d7cb44cb872911b5b46318772cf" Oct 10 09:56:25 crc kubenswrapper[4669]: I1010 09:56:25.474803 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3"} Oct 10 09:56:41 crc kubenswrapper[4669]: I1010 09:56:41.602152 4669 generic.go:334] "Generic (PLEG): container finished" podID="f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" containerID="c853cf72a7b476d87a0be0c8ee3bc29935cd6aa32fbd254442187226df57ed5e" exitCode=0 Oct 10 09:56:41 crc kubenswrapper[4669]: I1010 09:56:41.602242 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" event={"ID":"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e","Type":"ContainerDied","Data":"c853cf72a7b476d87a0be0c8ee3bc29935cd6aa32fbd254442187226df57ed5e"} Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.018896 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.118974 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.119212 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ceph\") pod \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.119353 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ssh-key\") pod \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.120225 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-metadata-combined-ca-bundle\") pod \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.120326 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8tpz\" (UniqueName: \"kubernetes.io/projected/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-kube-api-access-n8tpz\") pod \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.120431 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-nova-metadata-neutron-config-0\") pod \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.120602 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-inventory\") pod \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\" (UID: \"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e\") " Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.125327 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ceph" (OuterVolumeSpecName: "ceph") pod "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" (UID: "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.125357 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-kube-api-access-n8tpz" (OuterVolumeSpecName: "kube-api-access-n8tpz") pod "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" (UID: "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e"). InnerVolumeSpecName "kube-api-access-n8tpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.137960 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" (UID: "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.146127 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" (UID: "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.149443 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-inventory" (OuterVolumeSpecName: "inventory") pod "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" (UID: "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.150294 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" (UID: "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.153922 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" (UID: "f9d4c45c-5b27-4012-a0d0-72a1a2b1843e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.222380 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8tpz\" (UniqueName: \"kubernetes.io/projected/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-kube-api-access-n8tpz\") on node \"crc\" DevicePath \"\"" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.222419 4669 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.222430 4669 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.222440 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.222450 4669 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.222462 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.222472 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9d4c45c-5b27-4012-a0d0-72a1a2b1843e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.617843 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" event={"ID":"f9d4c45c-5b27-4012-a0d0-72a1a2b1843e","Type":"ContainerDied","Data":"e4e7e132f3e4d6eb012a9f69c47e402e3da2d432d20abe217b6d73b67fb17192"} Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.617887 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4e7e132f3e4d6eb012a9f69c47e402e3da2d432d20abe217b6d73b67fb17192" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.617892 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.720700 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62"] Oct 10 09:56:43 crc kubenswrapper[4669]: E1010 09:56:43.721125 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.721149 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.721384 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9d4c45c-5b27-4012-a0d0-72a1a2b1843e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.722187 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.725219 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.727894 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.727933 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.727894 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.727982 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.728242 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.731048 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62"] Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.838179 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.838265 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.838326 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.838454 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.838527 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.838637 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvj7b\" (UniqueName: \"kubernetes.io/projected/fbaa41c7-694f-4fda-aaf5-481188158475-kube-api-access-kvj7b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.940652 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvj7b\" (UniqueName: \"kubernetes.io/projected/fbaa41c7-694f-4fda-aaf5-481188158475-kube-api-access-kvj7b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.940753 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.940810 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.940857 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.940916 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.941182 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.944915 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.945136 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.945305 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.946335 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.946337 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:43 crc kubenswrapper[4669]: I1010 09:56:43.960138 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvj7b\" (UniqueName: \"kubernetes.io/projected/fbaa41c7-694f-4fda-aaf5-481188158475-kube-api-access-kvj7b\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-8kd62\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:44 crc kubenswrapper[4669]: I1010 09:56:44.083666 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 09:56:44 crc kubenswrapper[4669]: I1010 09:56:44.649374 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62"] Oct 10 09:56:45 crc kubenswrapper[4669]: I1010 09:56:45.633488 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" event={"ID":"fbaa41c7-694f-4fda-aaf5-481188158475","Type":"ContainerStarted","Data":"b2bd38e8f2f6bf51be12c7ba8d8e230b443f9d790c9c88dcc8b11381b6091e47"} Oct 10 09:56:45 crc kubenswrapper[4669]: I1010 09:56:45.633845 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" event={"ID":"fbaa41c7-694f-4fda-aaf5-481188158475","Type":"ContainerStarted","Data":"90d7d5c67e74e375477e9a1cc2b02214ff44d6ba2f6236dac4f7e2d561af1f7c"} Oct 10 09:57:13 crc kubenswrapper[4669]: I1010 09:57:13.895108 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" podStartSLOduration=30.250062766 podStartE2EDuration="30.895090421s" podCreationTimestamp="2025-10-10 09:56:43 +0000 UTC" firstStartedPulling="2025-10-10 09:56:44.668388579 +0000 UTC m=+2747.684407331" lastFinishedPulling="2025-10-10 09:56:45.313416254 +0000 UTC m=+2748.329434986" observedRunningTime="2025-10-10 09:56:45.652169267 +0000 UTC m=+2748.668188009" watchObservedRunningTime="2025-10-10 09:57:13.895090421 +0000 UTC m=+2776.911109173" Oct 10 09:57:13 crc kubenswrapper[4669]: I1010 09:57:13.899246 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tbm8k"] Oct 10 09:57:13 crc kubenswrapper[4669]: I1010 09:57:13.901742 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:13 crc kubenswrapper[4669]: I1010 09:57:13.921631 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbm8k"] Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.054051 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-catalog-content\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.054571 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8bh8\" (UniqueName: \"kubernetes.io/projected/0f63cee5-2f0e-40f3-bbdb-562da0605575-kube-api-access-r8bh8\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.054782 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-utilities\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.156872 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-catalog-content\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.157385 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8bh8\" (UniqueName: \"kubernetes.io/projected/0f63cee5-2f0e-40f3-bbdb-562da0605575-kube-api-access-r8bh8\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.157427 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-catalog-content\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.157786 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-utilities\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.158143 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-utilities\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.176272 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8bh8\" (UniqueName: \"kubernetes.io/projected/0f63cee5-2f0e-40f3-bbdb-562da0605575-kube-api-access-r8bh8\") pod \"redhat-marketplace-tbm8k\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.222119 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.693367 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbm8k"] Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.911988 4669 generic.go:334] "Generic (PLEG): container finished" podID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerID="31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72" exitCode=0 Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.912051 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbm8k" event={"ID":"0f63cee5-2f0e-40f3-bbdb-562da0605575","Type":"ContainerDied","Data":"31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72"} Oct 10 09:57:14 crc kubenswrapper[4669]: I1010 09:57:14.912253 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbm8k" event={"ID":"0f63cee5-2f0e-40f3-bbdb-562da0605575","Type":"ContainerStarted","Data":"489fa5006b154523c645db6a697595e9efefa7bf7869f0ec7a46fa88a1e0e524"} Oct 10 09:57:15 crc kubenswrapper[4669]: I1010 09:57:15.923095 4669 generic.go:334] "Generic (PLEG): container finished" podID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerID="5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39" exitCode=0 Oct 10 09:57:15 crc kubenswrapper[4669]: I1010 09:57:15.923189 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbm8k" event={"ID":"0f63cee5-2f0e-40f3-bbdb-562da0605575","Type":"ContainerDied","Data":"5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39"} Oct 10 09:57:16 crc kubenswrapper[4669]: I1010 09:57:16.936184 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbm8k" event={"ID":"0f63cee5-2f0e-40f3-bbdb-562da0605575","Type":"ContainerStarted","Data":"ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634"} Oct 10 09:57:16 crc kubenswrapper[4669]: I1010 09:57:16.956484 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tbm8k" podStartSLOduration=2.5207520900000002 podStartE2EDuration="3.956468369s" podCreationTimestamp="2025-10-10 09:57:13 +0000 UTC" firstStartedPulling="2025-10-10 09:57:14.914047985 +0000 UTC m=+2777.930066727" lastFinishedPulling="2025-10-10 09:57:16.349764274 +0000 UTC m=+2779.365783006" observedRunningTime="2025-10-10 09:57:16.954202436 +0000 UTC m=+2779.970221178" watchObservedRunningTime="2025-10-10 09:57:16.956468369 +0000 UTC m=+2779.972487111" Oct 10 09:57:24 crc kubenswrapper[4669]: I1010 09:57:24.222931 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:24 crc kubenswrapper[4669]: I1010 09:57:24.223307 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:24 crc kubenswrapper[4669]: I1010 09:57:24.276741 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:25 crc kubenswrapper[4669]: I1010 09:57:25.050631 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:25 crc kubenswrapper[4669]: I1010 09:57:25.096272 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbm8k"] Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.020286 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tbm8k" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="registry-server" containerID="cri-o://ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634" gracePeriod=2 Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.422406 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.621872 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-catalog-content\") pod \"0f63cee5-2f0e-40f3-bbdb-562da0605575\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.621926 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-utilities\") pod \"0f63cee5-2f0e-40f3-bbdb-562da0605575\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.622165 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8bh8\" (UniqueName: \"kubernetes.io/projected/0f63cee5-2f0e-40f3-bbdb-562da0605575-kube-api-access-r8bh8\") pod \"0f63cee5-2f0e-40f3-bbdb-562da0605575\" (UID: \"0f63cee5-2f0e-40f3-bbdb-562da0605575\") " Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.622933 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-utilities" (OuterVolumeSpecName: "utilities") pod "0f63cee5-2f0e-40f3-bbdb-562da0605575" (UID: "0f63cee5-2f0e-40f3-bbdb-562da0605575"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.641538 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f63cee5-2f0e-40f3-bbdb-562da0605575" (UID: "0f63cee5-2f0e-40f3-bbdb-562da0605575"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.641829 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f63cee5-2f0e-40f3-bbdb-562da0605575-kube-api-access-r8bh8" (OuterVolumeSpecName: "kube-api-access-r8bh8") pod "0f63cee5-2f0e-40f3-bbdb-562da0605575" (UID: "0f63cee5-2f0e-40f3-bbdb-562da0605575"). InnerVolumeSpecName "kube-api-access-r8bh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.725190 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8bh8\" (UniqueName: \"kubernetes.io/projected/0f63cee5-2f0e-40f3-bbdb-562da0605575-kube-api-access-r8bh8\") on node \"crc\" DevicePath \"\"" Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.725524 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 09:57:27 crc kubenswrapper[4669]: I1010 09:57:27.725664 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f63cee5-2f0e-40f3-bbdb-562da0605575-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.034943 4669 generic.go:334] "Generic (PLEG): container finished" podID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerID="ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634" exitCode=0 Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.034987 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbm8k" event={"ID":"0f63cee5-2f0e-40f3-bbdb-562da0605575","Type":"ContainerDied","Data":"ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634"} Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.035004 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbm8k" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.035022 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbm8k" event={"ID":"0f63cee5-2f0e-40f3-bbdb-562da0605575","Type":"ContainerDied","Data":"489fa5006b154523c645db6a697595e9efefa7bf7869f0ec7a46fa88a1e0e524"} Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.035044 4669 scope.go:117] "RemoveContainer" containerID="ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.060230 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbm8k"] Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.072773 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbm8k"] Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.086895 4669 scope.go:117] "RemoveContainer" containerID="5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.109238 4669 scope.go:117] "RemoveContainer" containerID="31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.141575 4669 scope.go:117] "RemoveContainer" containerID="ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634" Oct 10 09:57:28 crc kubenswrapper[4669]: E1010 09:57:28.142264 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634\": container with ID starting with ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634 not found: ID does not exist" containerID="ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.142330 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634"} err="failed to get container status \"ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634\": rpc error: code = NotFound desc = could not find container \"ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634\": container with ID starting with ec0cf3d53007c99416bc65767a3634a46dd81cfb3ea6a881cdbfa9ba16222634 not found: ID does not exist" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.142358 4669 scope.go:117] "RemoveContainer" containerID="5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39" Oct 10 09:57:28 crc kubenswrapper[4669]: E1010 09:57:28.142727 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39\": container with ID starting with 5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39 not found: ID does not exist" containerID="5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.142753 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39"} err="failed to get container status \"5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39\": rpc error: code = NotFound desc = could not find container \"5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39\": container with ID starting with 5506a4d041d64960703b10c65e1616a65e9d606ef1b270e5809a7fda1a751b39 not found: ID does not exist" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.142795 4669 scope.go:117] "RemoveContainer" containerID="31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72" Oct 10 09:57:28 crc kubenswrapper[4669]: E1010 09:57:28.143508 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72\": container with ID starting with 31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72 not found: ID does not exist" containerID="31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72" Oct 10 09:57:28 crc kubenswrapper[4669]: I1010 09:57:28.143533 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72"} err="failed to get container status \"31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72\": rpc error: code = NotFound desc = could not find container \"31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72\": container with ID starting with 31fa23e37eede9a01a7c9d2dc5ab2476d6772ffea144353df102966032aeda72 not found: ID does not exist" Oct 10 09:57:29 crc kubenswrapper[4669]: I1010 09:57:29.833626 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" path="/var/lib/kubelet/pods/0f63cee5-2f0e-40f3-bbdb-562da0605575/volumes" Oct 10 09:58:24 crc kubenswrapper[4669]: I1010 09:58:24.275741 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:58:24 crc kubenswrapper[4669]: I1010 09:58:24.276814 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:58:54 crc kubenswrapper[4669]: I1010 09:58:54.275035 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:58:54 crc kubenswrapper[4669]: I1010 09:58:54.275690 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:59:24 crc kubenswrapper[4669]: I1010 09:59:24.274849 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 09:59:24 crc kubenswrapper[4669]: I1010 09:59:24.275425 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 09:59:24 crc kubenswrapper[4669]: I1010 09:59:24.275487 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 09:59:24 crc kubenswrapper[4669]: I1010 09:59:24.276438 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 09:59:24 crc kubenswrapper[4669]: I1010 09:59:24.276700 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" gracePeriod=600 Oct 10 09:59:24 crc kubenswrapper[4669]: E1010 09:59:24.414275 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:59:25 crc kubenswrapper[4669]: I1010 09:59:25.111333 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" exitCode=0 Oct 10 09:59:25 crc kubenswrapper[4669]: I1010 09:59:25.111382 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3"} Oct 10 09:59:25 crc kubenswrapper[4669]: I1010 09:59:25.111418 4669 scope.go:117] "RemoveContainer" containerID="17f3dc9423181b5b82c8eb0219ab63c353115b847743715862f372bb06c2885e" Oct 10 09:59:25 crc kubenswrapper[4669]: I1010 09:59:25.114086 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 09:59:25 crc kubenswrapper[4669]: E1010 09:59:25.114670 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:59:35 crc kubenswrapper[4669]: I1010 09:59:35.795999 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 09:59:35 crc kubenswrapper[4669]: E1010 09:59:35.796921 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 09:59:50 crc kubenswrapper[4669]: I1010 09:59:50.795987 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 09:59:50 crc kubenswrapper[4669]: E1010 09:59:50.800467 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.175685 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm"] Oct 10 10:00:00 crc kubenswrapper[4669]: E1010 10:00:00.176861 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="extract-utilities" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.176879 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="extract-utilities" Oct 10 10:00:00 crc kubenswrapper[4669]: E1010 10:00:00.176908 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="extract-content" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.176915 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="extract-content" Oct 10 10:00:00 crc kubenswrapper[4669]: E1010 10:00:00.176938 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="registry-server" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.176946 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="registry-server" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.177148 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f63cee5-2f0e-40f3-bbdb-562da0605575" containerName="registry-server" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.177897 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.181532 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.181832 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.185816 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm"] Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.370732 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5vl8\" (UniqueName: \"kubernetes.io/projected/e07fd613-5421-4b91-8407-d1fff86a8dbc-kube-api-access-b5vl8\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.370941 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e07fd613-5421-4b91-8407-d1fff86a8dbc-secret-volume\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.371047 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e07fd613-5421-4b91-8407-d1fff86a8dbc-config-volume\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.472448 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5vl8\" (UniqueName: \"kubernetes.io/projected/e07fd613-5421-4b91-8407-d1fff86a8dbc-kube-api-access-b5vl8\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.472536 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e07fd613-5421-4b91-8407-d1fff86a8dbc-secret-volume\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.472622 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e07fd613-5421-4b91-8407-d1fff86a8dbc-config-volume\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.473707 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e07fd613-5421-4b91-8407-d1fff86a8dbc-config-volume\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.479608 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e07fd613-5421-4b91-8407-d1fff86a8dbc-secret-volume\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.495270 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5vl8\" (UniqueName: \"kubernetes.io/projected/e07fd613-5421-4b91-8407-d1fff86a8dbc-kube-api-access-b5vl8\") pod \"collect-profiles-29334840-t2hzm\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.506555 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:00 crc kubenswrapper[4669]: I1010 10:00:00.980940 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm"] Oct 10 10:00:01 crc kubenswrapper[4669]: I1010 10:00:01.449466 4669 generic.go:334] "Generic (PLEG): container finished" podID="e07fd613-5421-4b91-8407-d1fff86a8dbc" containerID="d555f05d05ade41f496d2db9c4676016e9435104ec4e2b55644550a9cec2b348" exitCode=0 Oct 10 10:00:01 crc kubenswrapper[4669]: I1010 10:00:01.449521 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" event={"ID":"e07fd613-5421-4b91-8407-d1fff86a8dbc","Type":"ContainerDied","Data":"d555f05d05ade41f496d2db9c4676016e9435104ec4e2b55644550a9cec2b348"} Oct 10 10:00:01 crc kubenswrapper[4669]: I1010 10:00:01.449765 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" event={"ID":"e07fd613-5421-4b91-8407-d1fff86a8dbc","Type":"ContainerStarted","Data":"b13d1f1ae8f04a056749e2f14a20c6264361a30031b69969d83b03f7903e3b4f"} Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.797809 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:00:02 crc kubenswrapper[4669]: E1010 10:00:02.798288 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.803555 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.913446 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e07fd613-5421-4b91-8407-d1fff86a8dbc-config-volume\") pod \"e07fd613-5421-4b91-8407-d1fff86a8dbc\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.913736 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5vl8\" (UniqueName: \"kubernetes.io/projected/e07fd613-5421-4b91-8407-d1fff86a8dbc-kube-api-access-b5vl8\") pod \"e07fd613-5421-4b91-8407-d1fff86a8dbc\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.913786 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e07fd613-5421-4b91-8407-d1fff86a8dbc-secret-volume\") pod \"e07fd613-5421-4b91-8407-d1fff86a8dbc\" (UID: \"e07fd613-5421-4b91-8407-d1fff86a8dbc\") " Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.914291 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e07fd613-5421-4b91-8407-d1fff86a8dbc-config-volume" (OuterVolumeSpecName: "config-volume") pod "e07fd613-5421-4b91-8407-d1fff86a8dbc" (UID: "e07fd613-5421-4b91-8407-d1fff86a8dbc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.915621 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e07fd613-5421-4b91-8407-d1fff86a8dbc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.919130 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e07fd613-5421-4b91-8407-d1fff86a8dbc-kube-api-access-b5vl8" (OuterVolumeSpecName: "kube-api-access-b5vl8") pod "e07fd613-5421-4b91-8407-d1fff86a8dbc" (UID: "e07fd613-5421-4b91-8407-d1fff86a8dbc"). InnerVolumeSpecName "kube-api-access-b5vl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:00:02 crc kubenswrapper[4669]: I1010 10:00:02.928713 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e07fd613-5421-4b91-8407-d1fff86a8dbc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e07fd613-5421-4b91-8407-d1fff86a8dbc" (UID: "e07fd613-5421-4b91-8407-d1fff86a8dbc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:00:03 crc kubenswrapper[4669]: I1010 10:00:03.017194 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5vl8\" (UniqueName: \"kubernetes.io/projected/e07fd613-5421-4b91-8407-d1fff86a8dbc-kube-api-access-b5vl8\") on node \"crc\" DevicePath \"\"" Oct 10 10:00:03 crc kubenswrapper[4669]: I1010 10:00:03.017228 4669 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e07fd613-5421-4b91-8407-d1fff86a8dbc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 10:00:03 crc kubenswrapper[4669]: I1010 10:00:03.466614 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" event={"ID":"e07fd613-5421-4b91-8407-d1fff86a8dbc","Type":"ContainerDied","Data":"b13d1f1ae8f04a056749e2f14a20c6264361a30031b69969d83b03f7903e3b4f"} Oct 10 10:00:03 crc kubenswrapper[4669]: I1010 10:00:03.466990 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b13d1f1ae8f04a056749e2f14a20c6264361a30031b69969d83b03f7903e3b4f" Oct 10 10:00:03 crc kubenswrapper[4669]: I1010 10:00:03.466684 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334840-t2hzm" Oct 10 10:00:03 crc kubenswrapper[4669]: I1010 10:00:03.894358 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf"] Oct 10 10:00:03 crc kubenswrapper[4669]: I1010 10:00:03.906009 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334795-f2xqf"] Oct 10 10:00:05 crc kubenswrapper[4669]: I1010 10:00:05.807819 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="973b9fdb-7b24-46a6-b351-25cbe8195797" path="/var/lib/kubelet/pods/973b9fdb-7b24-46a6-b351-25cbe8195797/volumes" Oct 10 10:00:12 crc kubenswrapper[4669]: I1010 10:00:12.925015 4669 scope.go:117] "RemoveContainer" containerID="49db09a03c7a401bd406b966d88aa62a27cf1a9e9625ca9d8bdbc064c7ac2f6a" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.467784 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-plw6f"] Oct 10 10:00:15 crc kubenswrapper[4669]: E1010 10:00:15.468451 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e07fd613-5421-4b91-8407-d1fff86a8dbc" containerName="collect-profiles" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.468464 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e07fd613-5421-4b91-8407-d1fff86a8dbc" containerName="collect-profiles" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.468666 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e07fd613-5421-4b91-8407-d1fff86a8dbc" containerName="collect-profiles" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.469884 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.484410 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-plw6f"] Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.659723 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-catalog-content\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.659806 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-utilities\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.659899 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck7l5\" (UniqueName: \"kubernetes.io/projected/2cb4b16d-de81-4b67-a23e-12f3139e151b-kube-api-access-ck7l5\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.761264 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck7l5\" (UniqueName: \"kubernetes.io/projected/2cb4b16d-de81-4b67-a23e-12f3139e151b-kube-api-access-ck7l5\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.761367 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-catalog-content\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.761408 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-utilities\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.761943 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-utilities\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.762015 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-catalog-content\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.782144 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck7l5\" (UniqueName: \"kubernetes.io/projected/2cb4b16d-de81-4b67-a23e-12f3139e151b-kube-api-access-ck7l5\") pod \"certified-operators-plw6f\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:15 crc kubenswrapper[4669]: I1010 10:00:15.792677 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:16 crc kubenswrapper[4669]: I1010 10:00:16.373475 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-plw6f"] Oct 10 10:00:16 crc kubenswrapper[4669]: I1010 10:00:16.576343 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plw6f" event={"ID":"2cb4b16d-de81-4b67-a23e-12f3139e151b","Type":"ContainerStarted","Data":"cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3"} Oct 10 10:00:16 crc kubenswrapper[4669]: I1010 10:00:16.576565 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plw6f" event={"ID":"2cb4b16d-de81-4b67-a23e-12f3139e151b","Type":"ContainerStarted","Data":"fa2e4b703234cd78755ee427497ad4165f32f14e5cc3a5b29236f9b694af16ed"} Oct 10 10:00:16 crc kubenswrapper[4669]: I1010 10:00:16.795020 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:00:16 crc kubenswrapper[4669]: E1010 10:00:16.795324 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:00:17 crc kubenswrapper[4669]: I1010 10:00:17.586712 4669 generic.go:334] "Generic (PLEG): container finished" podID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerID="cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3" exitCode=0 Oct 10 10:00:17 crc kubenswrapper[4669]: I1010 10:00:17.586756 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plw6f" event={"ID":"2cb4b16d-de81-4b67-a23e-12f3139e151b","Type":"ContainerDied","Data":"cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3"} Oct 10 10:00:17 crc kubenswrapper[4669]: I1010 10:00:17.590346 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 10:00:19 crc kubenswrapper[4669]: I1010 10:00:19.607046 4669 generic.go:334] "Generic (PLEG): container finished" podID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerID="f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f" exitCode=0 Oct 10 10:00:19 crc kubenswrapper[4669]: I1010 10:00:19.607561 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plw6f" event={"ID":"2cb4b16d-de81-4b67-a23e-12f3139e151b","Type":"ContainerDied","Data":"f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f"} Oct 10 10:00:20 crc kubenswrapper[4669]: I1010 10:00:20.617273 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plw6f" event={"ID":"2cb4b16d-de81-4b67-a23e-12f3139e151b","Type":"ContainerStarted","Data":"6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528"} Oct 10 10:00:20 crc kubenswrapper[4669]: I1010 10:00:20.641940 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-plw6f" podStartSLOduration=3.152670713 podStartE2EDuration="5.641920316s" podCreationTimestamp="2025-10-10 10:00:15 +0000 UTC" firstStartedPulling="2025-10-10 10:00:17.590055022 +0000 UTC m=+2960.606073774" lastFinishedPulling="2025-10-10 10:00:20.079304635 +0000 UTC m=+2963.095323377" observedRunningTime="2025-10-10 10:00:20.639759767 +0000 UTC m=+2963.655778649" watchObservedRunningTime="2025-10-10 10:00:20.641920316 +0000 UTC m=+2963.657939058" Oct 10 10:00:25 crc kubenswrapper[4669]: I1010 10:00:25.793434 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:25 crc kubenswrapper[4669]: I1010 10:00:25.794081 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:25 crc kubenswrapper[4669]: I1010 10:00:25.860639 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:26 crc kubenswrapper[4669]: I1010 10:00:26.730510 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:26 crc kubenswrapper[4669]: I1010 10:00:26.780688 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-plw6f"] Oct 10 10:00:27 crc kubenswrapper[4669]: I1010 10:00:27.803184 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:00:27 crc kubenswrapper[4669]: E1010 10:00:27.803499 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:00:28 crc kubenswrapper[4669]: I1010 10:00:28.701927 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-plw6f" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="registry-server" containerID="cri-o://6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528" gracePeriod=2 Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.135000 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.218548 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck7l5\" (UniqueName: \"kubernetes.io/projected/2cb4b16d-de81-4b67-a23e-12f3139e151b-kube-api-access-ck7l5\") pod \"2cb4b16d-de81-4b67-a23e-12f3139e151b\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.218690 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-catalog-content\") pod \"2cb4b16d-de81-4b67-a23e-12f3139e151b\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.218734 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-utilities\") pod \"2cb4b16d-de81-4b67-a23e-12f3139e151b\" (UID: \"2cb4b16d-de81-4b67-a23e-12f3139e151b\") " Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.219452 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-utilities" (OuterVolumeSpecName: "utilities") pod "2cb4b16d-de81-4b67-a23e-12f3139e151b" (UID: "2cb4b16d-de81-4b67-a23e-12f3139e151b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.224025 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb4b16d-de81-4b67-a23e-12f3139e151b-kube-api-access-ck7l5" (OuterVolumeSpecName: "kube-api-access-ck7l5") pod "2cb4b16d-de81-4b67-a23e-12f3139e151b" (UID: "2cb4b16d-de81-4b67-a23e-12f3139e151b"). InnerVolumeSpecName "kube-api-access-ck7l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.294396 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cb4b16d-de81-4b67-a23e-12f3139e151b" (UID: "2cb4b16d-de81-4b67-a23e-12f3139e151b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.321112 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.321155 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck7l5\" (UniqueName: \"kubernetes.io/projected/2cb4b16d-de81-4b67-a23e-12f3139e151b-kube-api-access-ck7l5\") on node \"crc\" DevicePath \"\"" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.321165 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cb4b16d-de81-4b67-a23e-12f3139e151b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.713521 4669 generic.go:334] "Generic (PLEG): container finished" podID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerID="6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528" exitCode=0 Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.713579 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plw6f" event={"ID":"2cb4b16d-de81-4b67-a23e-12f3139e151b","Type":"ContainerDied","Data":"6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528"} Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.713624 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plw6f" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.713646 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plw6f" event={"ID":"2cb4b16d-de81-4b67-a23e-12f3139e151b","Type":"ContainerDied","Data":"fa2e4b703234cd78755ee427497ad4165f32f14e5cc3a5b29236f9b694af16ed"} Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.713676 4669 scope.go:117] "RemoveContainer" containerID="6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.738877 4669 scope.go:117] "RemoveContainer" containerID="f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.776622 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-plw6f"] Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.780499 4669 scope.go:117] "RemoveContainer" containerID="cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.807889 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-plw6f"] Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.821818 4669 scope.go:117] "RemoveContainer" containerID="6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528" Oct 10 10:00:29 crc kubenswrapper[4669]: E1010 10:00:29.824476 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528\": container with ID starting with 6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528 not found: ID does not exist" containerID="6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.824528 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528"} err="failed to get container status \"6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528\": rpc error: code = NotFound desc = could not find container \"6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528\": container with ID starting with 6dbc6263809188c53ab62fe518ba5790101cd80449e0c34eec6fb3c5061af528 not found: ID does not exist" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.824563 4669 scope.go:117] "RemoveContainer" containerID="f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f" Oct 10 10:00:29 crc kubenswrapper[4669]: E1010 10:00:29.826353 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f\": container with ID starting with f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f not found: ID does not exist" containerID="f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.826389 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f"} err="failed to get container status \"f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f\": rpc error: code = NotFound desc = could not find container \"f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f\": container with ID starting with f7e4e485a23b288cf49df17d84f301dddc35b7af89a62b7e7cdc0687466bf88f not found: ID does not exist" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.826411 4669 scope.go:117] "RemoveContainer" containerID="cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3" Oct 10 10:00:29 crc kubenswrapper[4669]: E1010 10:00:29.829899 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3\": container with ID starting with cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3 not found: ID does not exist" containerID="cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3" Oct 10 10:00:29 crc kubenswrapper[4669]: I1010 10:00:29.829938 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3"} err="failed to get container status \"cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3\": rpc error: code = NotFound desc = could not find container \"cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3\": container with ID starting with cfb90593e2525ca30281be2c94ee728fe2e325f516862d1d990deb2eebd414b3 not found: ID does not exist" Oct 10 10:00:31 crc kubenswrapper[4669]: I1010 10:00:31.809651 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" path="/var/lib/kubelet/pods/2cb4b16d-de81-4b67-a23e-12f3139e151b/volumes" Oct 10 10:00:38 crc kubenswrapper[4669]: I1010 10:00:38.795116 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:00:38 crc kubenswrapper[4669]: E1010 10:00:38.796380 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:00:49 crc kubenswrapper[4669]: I1010 10:00:49.795524 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:00:49 crc kubenswrapper[4669]: E1010 10:00:49.796282 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.160239 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29334841-rb65h"] Oct 10 10:01:00 crc kubenswrapper[4669]: E1010 10:01:00.162376 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="extract-content" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.162426 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="extract-content" Oct 10 10:01:00 crc kubenswrapper[4669]: E1010 10:01:00.162447 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="extract-utilities" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.162454 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="extract-utilities" Oct 10 10:01:00 crc kubenswrapper[4669]: E1010 10:01:00.162482 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="registry-server" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.162493 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="registry-server" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.162785 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb4b16d-de81-4b67-a23e-12f3139e151b" containerName="registry-server" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.163805 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.178240 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29334841-rb65h"] Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.187227 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-combined-ca-bundle\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.187297 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-config-data\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.187325 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsdp4\" (UniqueName: \"kubernetes.io/projected/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-kube-api-access-lsdp4\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.187391 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-fernet-keys\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.288480 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-combined-ca-bundle\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.288530 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-config-data\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.288552 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsdp4\" (UniqueName: \"kubernetes.io/projected/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-kube-api-access-lsdp4\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.288614 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-fernet-keys\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.295258 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-fernet-keys\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.296274 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-combined-ca-bundle\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.306614 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-config-data\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.306792 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsdp4\" (UniqueName: \"kubernetes.io/projected/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-kube-api-access-lsdp4\") pod \"keystone-cron-29334841-rb65h\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:00 crc kubenswrapper[4669]: I1010 10:01:00.481742 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:01 crc kubenswrapper[4669]: I1010 10:01:01.152187 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29334841-rb65h"] Oct 10 10:01:01 crc kubenswrapper[4669]: I1010 10:01:01.797306 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:01:01 crc kubenswrapper[4669]: E1010 10:01:01.797880 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:01:01 crc kubenswrapper[4669]: I1010 10:01:01.995175 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29334841-rb65h" event={"ID":"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759","Type":"ContainerStarted","Data":"d2ed7cba1347d1b59f29b4c8bbe7d780bb9248090c2bb518282204d68220be67"} Oct 10 10:01:01 crc kubenswrapper[4669]: I1010 10:01:01.995228 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29334841-rb65h" event={"ID":"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759","Type":"ContainerStarted","Data":"fbb8d01f93b43b7a2f34676aec9fda4ef1d70d3bdf2744eb3fd2daca2392ee73"} Oct 10 10:01:02 crc kubenswrapper[4669]: I1010 10:01:02.020370 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29334841-rb65h" podStartSLOduration=2.020347787 podStartE2EDuration="2.020347787s" podCreationTimestamp="2025-10-10 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:01:02.016166682 +0000 UTC m=+3005.032185424" watchObservedRunningTime="2025-10-10 10:01:02.020347787 +0000 UTC m=+3005.036366529" Oct 10 10:01:05 crc kubenswrapper[4669]: I1010 10:01:05.021805 4669 generic.go:334] "Generic (PLEG): container finished" podID="d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" containerID="d2ed7cba1347d1b59f29b4c8bbe7d780bb9248090c2bb518282204d68220be67" exitCode=0 Oct 10 10:01:05 crc kubenswrapper[4669]: I1010 10:01:05.021876 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29334841-rb65h" event={"ID":"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759","Type":"ContainerDied","Data":"d2ed7cba1347d1b59f29b4c8bbe7d780bb9248090c2bb518282204d68220be67"} Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.329224 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.395647 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-combined-ca-bundle\") pod \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.395722 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-config-data\") pod \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.395810 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsdp4\" (UniqueName: \"kubernetes.io/projected/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-kube-api-access-lsdp4\") pod \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.395866 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-fernet-keys\") pod \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\" (UID: \"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759\") " Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.403877 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" (UID: "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.403898 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-kube-api-access-lsdp4" (OuterVolumeSpecName: "kube-api-access-lsdp4") pod "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" (UID: "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759"). InnerVolumeSpecName "kube-api-access-lsdp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.439970 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" (UID: "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.452525 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-config-data" (OuterVolumeSpecName: "config-data") pod "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" (UID: "d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.498043 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.498076 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.498086 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsdp4\" (UniqueName: \"kubernetes.io/projected/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-kube-api-access-lsdp4\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:06 crc kubenswrapper[4669]: I1010 10:01:06.498097 4669 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:07 crc kubenswrapper[4669]: I1010 10:01:07.039728 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29334841-rb65h" event={"ID":"d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759","Type":"ContainerDied","Data":"fbb8d01f93b43b7a2f34676aec9fda4ef1d70d3bdf2744eb3fd2daca2392ee73"} Oct 10 10:01:07 crc kubenswrapper[4669]: I1010 10:01:07.039776 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbb8d01f93b43b7a2f34676aec9fda4ef1d70d3bdf2744eb3fd2daca2392ee73" Oct 10 10:01:07 crc kubenswrapper[4669]: I1010 10:01:07.039860 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29334841-rb65h" Oct 10 10:01:13 crc kubenswrapper[4669]: I1010 10:01:13.795424 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:01:13 crc kubenswrapper[4669]: E1010 10:01:13.796055 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:01:26 crc kubenswrapper[4669]: I1010 10:01:26.795267 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:01:26 crc kubenswrapper[4669]: E1010 10:01:26.796083 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:01:36 crc kubenswrapper[4669]: I1010 10:01:36.323870 4669 generic.go:334] "Generic (PLEG): container finished" podID="fbaa41c7-694f-4fda-aaf5-481188158475" containerID="b2bd38e8f2f6bf51be12c7ba8d8e230b443f9d790c9c88dcc8b11381b6091e47" exitCode=0 Oct 10 10:01:36 crc kubenswrapper[4669]: I1010 10:01:36.323957 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" event={"ID":"fbaa41c7-694f-4fda-aaf5-481188158475","Type":"ContainerDied","Data":"b2bd38e8f2f6bf51be12c7ba8d8e230b443f9d790c9c88dcc8b11381b6091e47"} Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.779199 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.797409 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:01:37 crc kubenswrapper[4669]: E1010 10:01:37.797633 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.878472 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvj7b\" (UniqueName: \"kubernetes.io/projected/fbaa41c7-694f-4fda-aaf5-481188158475-kube-api-access-kvj7b\") pod \"fbaa41c7-694f-4fda-aaf5-481188158475\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.878520 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ceph\") pod \"fbaa41c7-694f-4fda-aaf5-481188158475\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.878543 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ssh-key\") pod \"fbaa41c7-694f-4fda-aaf5-481188158475\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.878564 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-secret-0\") pod \"fbaa41c7-694f-4fda-aaf5-481188158475\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.878608 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-inventory\") pod \"fbaa41c7-694f-4fda-aaf5-481188158475\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.878849 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-combined-ca-bundle\") pod \"fbaa41c7-694f-4fda-aaf5-481188158475\" (UID: \"fbaa41c7-694f-4fda-aaf5-481188158475\") " Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.884520 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fbaa41c7-694f-4fda-aaf5-481188158475" (UID: "fbaa41c7-694f-4fda-aaf5-481188158475"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.884867 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ceph" (OuterVolumeSpecName: "ceph") pod "fbaa41c7-694f-4fda-aaf5-481188158475" (UID: "fbaa41c7-694f-4fda-aaf5-481188158475"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.885896 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbaa41c7-694f-4fda-aaf5-481188158475-kube-api-access-kvj7b" (OuterVolumeSpecName: "kube-api-access-kvj7b") pod "fbaa41c7-694f-4fda-aaf5-481188158475" (UID: "fbaa41c7-694f-4fda-aaf5-481188158475"). InnerVolumeSpecName "kube-api-access-kvj7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.909802 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "fbaa41c7-694f-4fda-aaf5-481188158475" (UID: "fbaa41c7-694f-4fda-aaf5-481188158475"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.910213 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-inventory" (OuterVolumeSpecName: "inventory") pod "fbaa41c7-694f-4fda-aaf5-481188158475" (UID: "fbaa41c7-694f-4fda-aaf5-481188158475"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.923273 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fbaa41c7-694f-4fda-aaf5-481188158475" (UID: "fbaa41c7-694f-4fda-aaf5-481188158475"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.981814 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvj7b\" (UniqueName: \"kubernetes.io/projected/fbaa41c7-694f-4fda-aaf5-481188158475-kube-api-access-kvj7b\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.981840 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.981848 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.981859 4669 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.981869 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:37 crc kubenswrapper[4669]: I1010 10:01:37.981879 4669 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa41c7-694f-4fda-aaf5-481188158475-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.348925 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" event={"ID":"fbaa41c7-694f-4fda-aaf5-481188158475","Type":"ContainerDied","Data":"90d7d5c67e74e375477e9a1cc2b02214ff44d6ba2f6236dac4f7e2d561af1f7c"} Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.349207 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90d7d5c67e74e375477e9a1cc2b02214ff44d6ba2f6236dac4f7e2d561af1f7c" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.349002 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-8kd62" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.477698 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6"] Oct 10 10:01:38 crc kubenswrapper[4669]: E1010 10:01:38.478142 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" containerName="keystone-cron" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.478162 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" containerName="keystone-cron" Oct 10 10:01:38 crc kubenswrapper[4669]: E1010 10:01:38.478180 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbaa41c7-694f-4fda-aaf5-481188158475" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.478189 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbaa41c7-694f-4fda-aaf5-481188158475" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.478386 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbaa41c7-694f-4fda-aaf5-481188158475" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.478420 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759" containerName="keystone-cron" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.479553 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.483888 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.484326 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.484516 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2np6p" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.491785 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.492927 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.493199 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.493347 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.493466 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.493576 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.498165 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6"] Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593266 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593331 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593368 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593388 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593405 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593421 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52twn\" (UniqueName: \"kubernetes.io/projected/4933ae0b-dda6-44e1-a551-730e803bb7d5-kube-api-access-52twn\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593445 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593473 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593516 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593540 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.593615 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.696194 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.696307 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.696363 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.696406 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.697014 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.697071 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.697098 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52twn\" (UniqueName: \"kubernetes.io/projected/4933ae0b-dda6-44e1-a551-730e803bb7d5-kube-api-access-52twn\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.697132 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.697180 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.697259 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.697300 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.698149 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.699171 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.701130 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.701555 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.702143 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.702514 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.703448 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.705137 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.705176 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.709100 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.721636 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52twn\" (UniqueName: \"kubernetes.io/projected/4933ae0b-dda6-44e1-a551-730e803bb7d5-kube-api-access-52twn\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:38 crc kubenswrapper[4669]: I1010 10:01:38.803332 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:01:39 crc kubenswrapper[4669]: I1010 10:01:39.385210 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6"] Oct 10 10:01:40 crc kubenswrapper[4669]: I1010 10:01:40.395898 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" event={"ID":"4933ae0b-dda6-44e1-a551-730e803bb7d5","Type":"ContainerStarted","Data":"d68afb6cb3f20a274bc1746ca98b44c29a81cf939e50672734ceba4c27f73ae2"} Oct 10 10:01:40 crc kubenswrapper[4669]: I1010 10:01:40.396544 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" event={"ID":"4933ae0b-dda6-44e1-a551-730e803bb7d5","Type":"ContainerStarted","Data":"d203367db94e233f653183cc6f027da4441d539749984e0e521fb3fa264aee9e"} Oct 10 10:01:40 crc kubenswrapper[4669]: I1010 10:01:40.425894 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" podStartSLOduration=1.786571014 podStartE2EDuration="2.425870368s" podCreationTimestamp="2025-10-10 10:01:38 +0000 UTC" firstStartedPulling="2025-10-10 10:01:39.39333531 +0000 UTC m=+3042.409354062" lastFinishedPulling="2025-10-10 10:01:40.032634644 +0000 UTC m=+3043.048653416" observedRunningTime="2025-10-10 10:01:40.421793058 +0000 UTC m=+3043.437811820" watchObservedRunningTime="2025-10-10 10:01:40.425870368 +0000 UTC m=+3043.441889130" Oct 10 10:01:51 crc kubenswrapper[4669]: I1010 10:01:51.796047 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:01:51 crc kubenswrapper[4669]: E1010 10:01:51.797076 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:02:02 crc kubenswrapper[4669]: I1010 10:02:02.796109 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:02:02 crc kubenswrapper[4669]: E1010 10:02:02.797342 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:02:16 crc kubenswrapper[4669]: I1010 10:02:16.795969 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:02:16 crc kubenswrapper[4669]: E1010 10:02:16.797248 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:02:31 crc kubenswrapper[4669]: I1010 10:02:31.801077 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:02:31 crc kubenswrapper[4669]: E1010 10:02:31.802255 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:02:43 crc kubenswrapper[4669]: I1010 10:02:43.796141 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:02:43 crc kubenswrapper[4669]: E1010 10:02:43.797000 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:02:56 crc kubenswrapper[4669]: I1010 10:02:56.795187 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:02:56 crc kubenswrapper[4669]: E1010 10:02:56.795977 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:03:11 crc kubenswrapper[4669]: I1010 10:03:11.795846 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:03:11 crc kubenswrapper[4669]: E1010 10:03:11.796722 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.172963 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vzcvw"] Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.176715 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.183265 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzcvw"] Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.347826 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slgf4\" (UniqueName: \"kubernetes.io/projected/12c8a77b-d3fd-4bef-9465-84b05e7a7032-kube-api-access-slgf4\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.348169 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-utilities\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.348302 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-catalog-content\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.449482 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-utilities\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.449648 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-catalog-content\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.449740 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slgf4\" (UniqueName: \"kubernetes.io/projected/12c8a77b-d3fd-4bef-9465-84b05e7a7032-kube-api-access-slgf4\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.450636 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-utilities\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.450910 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-catalog-content\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.472098 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slgf4\" (UniqueName: \"kubernetes.io/projected/12c8a77b-d3fd-4bef-9465-84b05e7a7032-kube-api-access-slgf4\") pod \"redhat-operators-vzcvw\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:21 crc kubenswrapper[4669]: I1010 10:03:21.522685 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:22 crc kubenswrapper[4669]: I1010 10:03:22.058014 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzcvw"] Oct 10 10:03:22 crc kubenswrapper[4669]: I1010 10:03:22.272622 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcvw" event={"ID":"12c8a77b-d3fd-4bef-9465-84b05e7a7032","Type":"ContainerStarted","Data":"46022b68b73b6a80f6357c1d129dfd5e38a19ac3e3d4026258468da1116fc9ea"} Oct 10 10:03:23 crc kubenswrapper[4669]: I1010 10:03:23.282211 4669 generic.go:334] "Generic (PLEG): container finished" podID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerID="4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a" exitCode=0 Oct 10 10:03:23 crc kubenswrapper[4669]: I1010 10:03:23.282250 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcvw" event={"ID":"12c8a77b-d3fd-4bef-9465-84b05e7a7032","Type":"ContainerDied","Data":"4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a"} Oct 10 10:03:23 crc kubenswrapper[4669]: I1010 10:03:23.795850 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:03:23 crc kubenswrapper[4669]: E1010 10:03:23.796367 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:03:25 crc kubenswrapper[4669]: I1010 10:03:25.304473 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcvw" event={"ID":"12c8a77b-d3fd-4bef-9465-84b05e7a7032","Type":"ContainerStarted","Data":"a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b"} Oct 10 10:03:28 crc kubenswrapper[4669]: I1010 10:03:28.329200 4669 generic.go:334] "Generic (PLEG): container finished" podID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerID="a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b" exitCode=0 Oct 10 10:03:28 crc kubenswrapper[4669]: I1010 10:03:28.329277 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcvw" event={"ID":"12c8a77b-d3fd-4bef-9465-84b05e7a7032","Type":"ContainerDied","Data":"a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b"} Oct 10 10:03:29 crc kubenswrapper[4669]: I1010 10:03:29.341909 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcvw" event={"ID":"12c8a77b-d3fd-4bef-9465-84b05e7a7032","Type":"ContainerStarted","Data":"1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794"} Oct 10 10:03:29 crc kubenswrapper[4669]: I1010 10:03:29.375619 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vzcvw" podStartSLOduration=2.806748291 podStartE2EDuration="8.375570929s" podCreationTimestamp="2025-10-10 10:03:21 +0000 UTC" firstStartedPulling="2025-10-10 10:03:23.283732526 +0000 UTC m=+3146.299751268" lastFinishedPulling="2025-10-10 10:03:28.852555164 +0000 UTC m=+3151.868573906" observedRunningTime="2025-10-10 10:03:29.36779512 +0000 UTC m=+3152.383813862" watchObservedRunningTime="2025-10-10 10:03:29.375570929 +0000 UTC m=+3152.391589681" Oct 10 10:03:31 crc kubenswrapper[4669]: I1010 10:03:31.523281 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:31 crc kubenswrapper[4669]: I1010 10:03:31.523917 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:32 crc kubenswrapper[4669]: I1010 10:03:32.580689 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vzcvw" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="registry-server" probeResult="failure" output=< Oct 10 10:03:32 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 10:03:32 crc kubenswrapper[4669]: > Oct 10 10:03:35 crc kubenswrapper[4669]: I1010 10:03:35.796120 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:03:35 crc kubenswrapper[4669]: E1010 10:03:35.797048 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:03:41 crc kubenswrapper[4669]: I1010 10:03:41.580646 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:41 crc kubenswrapper[4669]: I1010 10:03:41.635234 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:41 crc kubenswrapper[4669]: I1010 10:03:41.817419 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzcvw"] Oct 10 10:03:43 crc kubenswrapper[4669]: I1010 10:03:43.469024 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vzcvw" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="registry-server" containerID="cri-o://1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794" gracePeriod=2 Oct 10 10:03:43 crc kubenswrapper[4669]: I1010 10:03:43.913339 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.059027 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slgf4\" (UniqueName: \"kubernetes.io/projected/12c8a77b-d3fd-4bef-9465-84b05e7a7032-kube-api-access-slgf4\") pod \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.059197 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-catalog-content\") pod \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.059320 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-utilities\") pod \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\" (UID: \"12c8a77b-d3fd-4bef-9465-84b05e7a7032\") " Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.060521 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-utilities" (OuterVolumeSpecName: "utilities") pod "12c8a77b-d3fd-4bef-9465-84b05e7a7032" (UID: "12c8a77b-d3fd-4bef-9465-84b05e7a7032"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.060967 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.067170 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c8a77b-d3fd-4bef-9465-84b05e7a7032-kube-api-access-slgf4" (OuterVolumeSpecName: "kube-api-access-slgf4") pod "12c8a77b-d3fd-4bef-9465-84b05e7a7032" (UID: "12c8a77b-d3fd-4bef-9465-84b05e7a7032"). InnerVolumeSpecName "kube-api-access-slgf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.162168 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slgf4\" (UniqueName: \"kubernetes.io/projected/12c8a77b-d3fd-4bef-9465-84b05e7a7032-kube-api-access-slgf4\") on node \"crc\" DevicePath \"\"" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.173337 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12c8a77b-d3fd-4bef-9465-84b05e7a7032" (UID: "12c8a77b-d3fd-4bef-9465-84b05e7a7032"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.263382 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c8a77b-d3fd-4bef-9465-84b05e7a7032-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.481371 4669 generic.go:334] "Generic (PLEG): container finished" podID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerID="1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794" exitCode=0 Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.481419 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcvw" event={"ID":"12c8a77b-d3fd-4bef-9465-84b05e7a7032","Type":"ContainerDied","Data":"1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794"} Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.481450 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzcvw" event={"ID":"12c8a77b-d3fd-4bef-9465-84b05e7a7032","Type":"ContainerDied","Data":"46022b68b73b6a80f6357c1d129dfd5e38a19ac3e3d4026258468da1116fc9ea"} Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.481446 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzcvw" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.481468 4669 scope.go:117] "RemoveContainer" containerID="1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.504529 4669 scope.go:117] "RemoveContainer" containerID="a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.522119 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzcvw"] Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.530323 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vzcvw"] Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.545251 4669 scope.go:117] "RemoveContainer" containerID="4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.578832 4669 scope.go:117] "RemoveContainer" containerID="1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794" Oct 10 10:03:44 crc kubenswrapper[4669]: E1010 10:03:44.580143 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794\": container with ID starting with 1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794 not found: ID does not exist" containerID="1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.580186 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794"} err="failed to get container status \"1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794\": rpc error: code = NotFound desc = could not find container \"1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794\": container with ID starting with 1c431c55b9d334bb439ff1f3408f6d976d34fa7471bce261cae033cd2b1e8794 not found: ID does not exist" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.580214 4669 scope.go:117] "RemoveContainer" containerID="a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b" Oct 10 10:03:44 crc kubenswrapper[4669]: E1010 10:03:44.580617 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b\": container with ID starting with a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b not found: ID does not exist" containerID="a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.580646 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b"} err="failed to get container status \"a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b\": rpc error: code = NotFound desc = could not find container \"a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b\": container with ID starting with a401b8c71de1cad5a15313cdf78571de024cb27191c4fa05ed45ffeb1459fc7b not found: ID does not exist" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.580666 4669 scope.go:117] "RemoveContainer" containerID="4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a" Oct 10 10:03:44 crc kubenswrapper[4669]: E1010 10:03:44.581280 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a\": container with ID starting with 4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a not found: ID does not exist" containerID="4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a" Oct 10 10:03:44 crc kubenswrapper[4669]: I1010 10:03:44.581313 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a"} err="failed to get container status \"4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a\": rpc error: code = NotFound desc = could not find container \"4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a\": container with ID starting with 4b08ed3f2ce2f4f998f95cb8a026e224705dee24c18cc72628536e211d439f2a not found: ID does not exist" Oct 10 10:03:45 crc kubenswrapper[4669]: I1010 10:03:45.809622 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" path="/var/lib/kubelet/pods/12c8a77b-d3fd-4bef-9465-84b05e7a7032/volumes" Oct 10 10:03:50 crc kubenswrapper[4669]: I1010 10:03:50.796106 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:03:50 crc kubenswrapper[4669]: E1010 10:03:50.797229 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:04:02 crc kubenswrapper[4669]: I1010 10:04:02.796386 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:04:02 crc kubenswrapper[4669]: E1010 10:04:02.797674 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:04:15 crc kubenswrapper[4669]: I1010 10:04:15.796501 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:04:15 crc kubenswrapper[4669]: E1010 10:04:15.797262 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:04:27 crc kubenswrapper[4669]: I1010 10:04:27.803526 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:04:28 crc kubenswrapper[4669]: I1010 10:04:28.896884 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"b1f7630f1b3eab7f0d934e2b66664e0f5dce72b53d2fc00baa4932a36d925d92"} Oct 10 10:05:04 crc kubenswrapper[4669]: I1010 10:05:04.853211 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j9xdg"] Oct 10 10:05:04 crc kubenswrapper[4669]: E1010 10:05:04.855084 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="extract-content" Oct 10 10:05:04 crc kubenswrapper[4669]: I1010 10:05:04.855171 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="extract-content" Oct 10 10:05:04 crc kubenswrapper[4669]: E1010 10:05:04.855268 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="extract-utilities" Oct 10 10:05:04 crc kubenswrapper[4669]: I1010 10:05:04.855331 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="extract-utilities" Oct 10 10:05:04 crc kubenswrapper[4669]: E1010 10:05:04.855396 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="registry-server" Oct 10 10:05:04 crc kubenswrapper[4669]: I1010 10:05:04.855457 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="registry-server" Oct 10 10:05:04 crc kubenswrapper[4669]: I1010 10:05:04.855723 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c8a77b-d3fd-4bef-9465-84b05e7a7032" containerName="registry-server" Oct 10 10:05:04 crc kubenswrapper[4669]: I1010 10:05:04.857062 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:04 crc kubenswrapper[4669]: I1010 10:05:04.862451 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9xdg"] Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.001640 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-utilities\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.001974 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54jdc\" (UniqueName: \"kubernetes.io/projected/af3d6264-2366-49fa-9564-70306d79c5c8-kube-api-access-54jdc\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.002063 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-catalog-content\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.103620 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54jdc\" (UniqueName: \"kubernetes.io/projected/af3d6264-2366-49fa-9564-70306d79c5c8-kube-api-access-54jdc\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.103936 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-catalog-content\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.104206 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-utilities\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.104481 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-catalog-content\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.104631 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-utilities\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.126408 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54jdc\" (UniqueName: \"kubernetes.io/projected/af3d6264-2366-49fa-9564-70306d79c5c8-kube-api-access-54jdc\") pod \"community-operators-j9xdg\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.191199 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:05 crc kubenswrapper[4669]: I1010 10:05:05.766959 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9xdg"] Oct 10 10:05:06 crc kubenswrapper[4669]: I1010 10:05:06.215279 4669 generic.go:334] "Generic (PLEG): container finished" podID="af3d6264-2366-49fa-9564-70306d79c5c8" containerID="39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4" exitCode=0 Oct 10 10:05:06 crc kubenswrapper[4669]: I1010 10:05:06.215317 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9xdg" event={"ID":"af3d6264-2366-49fa-9564-70306d79c5c8","Type":"ContainerDied","Data":"39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4"} Oct 10 10:05:06 crc kubenswrapper[4669]: I1010 10:05:06.215370 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9xdg" event={"ID":"af3d6264-2366-49fa-9564-70306d79c5c8","Type":"ContainerStarted","Data":"686a6b6ce11daffaf6e748ea57cbbada4bde8bd2165be2fb116145d4d1d5a90c"} Oct 10 10:05:08 crc kubenswrapper[4669]: I1010 10:05:08.237292 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9xdg" event={"ID":"af3d6264-2366-49fa-9564-70306d79c5c8","Type":"ContainerStarted","Data":"4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b"} Oct 10 10:05:09 crc kubenswrapper[4669]: I1010 10:05:09.255028 4669 generic.go:334] "Generic (PLEG): container finished" podID="af3d6264-2366-49fa-9564-70306d79c5c8" containerID="4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b" exitCode=0 Oct 10 10:05:09 crc kubenswrapper[4669]: I1010 10:05:09.255088 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9xdg" event={"ID":"af3d6264-2366-49fa-9564-70306d79c5c8","Type":"ContainerDied","Data":"4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b"} Oct 10 10:05:10 crc kubenswrapper[4669]: I1010 10:05:10.272516 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9xdg" event={"ID":"af3d6264-2366-49fa-9564-70306d79c5c8","Type":"ContainerStarted","Data":"fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95"} Oct 10 10:05:10 crc kubenswrapper[4669]: I1010 10:05:10.297096 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j9xdg" podStartSLOduration=2.716751855 podStartE2EDuration="6.297072968s" podCreationTimestamp="2025-10-10 10:05:04 +0000 UTC" firstStartedPulling="2025-10-10 10:05:06.226672686 +0000 UTC m=+3249.242691428" lastFinishedPulling="2025-10-10 10:05:09.806993799 +0000 UTC m=+3252.823012541" observedRunningTime="2025-10-10 10:05:10.290306661 +0000 UTC m=+3253.306325403" watchObservedRunningTime="2025-10-10 10:05:10.297072968 +0000 UTC m=+3253.313091710" Oct 10 10:05:15 crc kubenswrapper[4669]: I1010 10:05:15.191312 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:15 crc kubenswrapper[4669]: I1010 10:05:15.192182 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:15 crc kubenswrapper[4669]: I1010 10:05:15.274284 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:15 crc kubenswrapper[4669]: I1010 10:05:15.362212 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:16 crc kubenswrapper[4669]: I1010 10:05:16.442079 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9xdg"] Oct 10 10:05:17 crc kubenswrapper[4669]: I1010 10:05:17.334921 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j9xdg" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="registry-server" containerID="cri-o://fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95" gracePeriod=2 Oct 10 10:05:17 crc kubenswrapper[4669]: I1010 10:05:17.816765 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:17 crc kubenswrapper[4669]: I1010 10:05:17.954867 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54jdc\" (UniqueName: \"kubernetes.io/projected/af3d6264-2366-49fa-9564-70306d79c5c8-kube-api-access-54jdc\") pod \"af3d6264-2366-49fa-9564-70306d79c5c8\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " Oct 10 10:05:17 crc kubenswrapper[4669]: I1010 10:05:17.955029 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-catalog-content\") pod \"af3d6264-2366-49fa-9564-70306d79c5c8\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " Oct 10 10:05:17 crc kubenswrapper[4669]: I1010 10:05:17.955069 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-utilities\") pod \"af3d6264-2366-49fa-9564-70306d79c5c8\" (UID: \"af3d6264-2366-49fa-9564-70306d79c5c8\") " Oct 10 10:05:17 crc kubenswrapper[4669]: I1010 10:05:17.956163 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-utilities" (OuterVolumeSpecName: "utilities") pod "af3d6264-2366-49fa-9564-70306d79c5c8" (UID: "af3d6264-2366-49fa-9564-70306d79c5c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:05:17 crc kubenswrapper[4669]: I1010 10:05:17.960355 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af3d6264-2366-49fa-9564-70306d79c5c8-kube-api-access-54jdc" (OuterVolumeSpecName: "kube-api-access-54jdc") pod "af3d6264-2366-49fa-9564-70306d79c5c8" (UID: "af3d6264-2366-49fa-9564-70306d79c5c8"). InnerVolumeSpecName "kube-api-access-54jdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.003142 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af3d6264-2366-49fa-9564-70306d79c5c8" (UID: "af3d6264-2366-49fa-9564-70306d79c5c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.058368 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.058437 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54jdc\" (UniqueName: \"kubernetes.io/projected/af3d6264-2366-49fa-9564-70306d79c5c8-kube-api-access-54jdc\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.058461 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af3d6264-2366-49fa-9564-70306d79c5c8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.347991 4669 generic.go:334] "Generic (PLEG): container finished" podID="af3d6264-2366-49fa-9564-70306d79c5c8" containerID="fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95" exitCode=0 Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.348063 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9xdg" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.348066 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9xdg" event={"ID":"af3d6264-2366-49fa-9564-70306d79c5c8","Type":"ContainerDied","Data":"fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95"} Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.348133 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9xdg" event={"ID":"af3d6264-2366-49fa-9564-70306d79c5c8","Type":"ContainerDied","Data":"686a6b6ce11daffaf6e748ea57cbbada4bde8bd2165be2fb116145d4d1d5a90c"} Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.348154 4669 scope.go:117] "RemoveContainer" containerID="fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.394616 4669 scope.go:117] "RemoveContainer" containerID="4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.402155 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9xdg"] Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.411507 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j9xdg"] Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.430948 4669 scope.go:117] "RemoveContainer" containerID="39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.476991 4669 scope.go:117] "RemoveContainer" containerID="fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95" Oct 10 10:05:18 crc kubenswrapper[4669]: E1010 10:05:18.477463 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95\": container with ID starting with fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95 not found: ID does not exist" containerID="fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.477510 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95"} err="failed to get container status \"fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95\": rpc error: code = NotFound desc = could not find container \"fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95\": container with ID starting with fd8d7c62b4ba55a98d23853d5efd7229b3de8c0346fee0c4664f46e53a969d95 not found: ID does not exist" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.477536 4669 scope.go:117] "RemoveContainer" containerID="4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b" Oct 10 10:05:18 crc kubenswrapper[4669]: E1010 10:05:18.478501 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b\": container with ID starting with 4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b not found: ID does not exist" containerID="4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.478530 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b"} err="failed to get container status \"4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b\": rpc error: code = NotFound desc = could not find container \"4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b\": container with ID starting with 4f0bb157cec0c8d21f18aa2bdb60947914e585850dd1c5e25a3dd5ed87e5db6b not found: ID does not exist" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.478549 4669 scope.go:117] "RemoveContainer" containerID="39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4" Oct 10 10:05:18 crc kubenswrapper[4669]: E1010 10:05:18.478902 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4\": container with ID starting with 39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4 not found: ID does not exist" containerID="39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4" Oct 10 10:05:18 crc kubenswrapper[4669]: I1010 10:05:18.478934 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4"} err="failed to get container status \"39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4\": rpc error: code = NotFound desc = could not find container \"39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4\": container with ID starting with 39fa4c05a59d1071326650f655ec9f32f74d0865b2561ca3971a9ba294f87da4 not found: ID does not exist" Oct 10 10:05:19 crc kubenswrapper[4669]: I1010 10:05:19.807934 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" path="/var/lib/kubelet/pods/af3d6264-2366-49fa-9564-70306d79c5c8/volumes" Oct 10 10:05:46 crc kubenswrapper[4669]: I1010 10:05:46.634965 4669 generic.go:334] "Generic (PLEG): container finished" podID="4933ae0b-dda6-44e1-a551-730e803bb7d5" containerID="d68afb6cb3f20a274bc1746ca98b44c29a81cf939e50672734ceba4c27f73ae2" exitCode=0 Oct 10 10:05:46 crc kubenswrapper[4669]: I1010 10:05:46.636445 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" event={"ID":"4933ae0b-dda6-44e1-a551-730e803bb7d5","Type":"ContainerDied","Data":"d68afb6cb3f20a274bc1746ca98b44c29a81cf939e50672734ceba4c27f73ae2"} Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.114898 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258444 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-0\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258632 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ssh-key\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258718 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-1\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258799 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-1\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258864 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-0\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258892 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52twn\" (UniqueName: \"kubernetes.io/projected/4933ae0b-dda6-44e1-a551-730e803bb7d5-kube-api-access-52twn\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258925 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.258966 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-extra-config-0\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.259006 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph-nova-0\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.259050 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-inventory\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.259082 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-custom-ceph-combined-ca-bundle\") pod \"4933ae0b-dda6-44e1-a551-730e803bb7d5\" (UID: \"4933ae0b-dda6-44e1-a551-730e803bb7d5\") " Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.264372 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.264554 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4933ae0b-dda6-44e1-a551-730e803bb7d5-kube-api-access-52twn" (OuterVolumeSpecName: "kube-api-access-52twn") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "kube-api-access-52twn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.266935 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph" (OuterVolumeSpecName: "ceph") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.287643 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.291154 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.292773 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.293825 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.293839 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.313823 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.317278 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.317558 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-inventory" (OuterVolumeSpecName: "inventory") pod "4933ae0b-dda6-44e1-a551-730e803bb7d5" (UID: "4933ae0b-dda6-44e1-a551-730e803bb7d5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361565 4669 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361606 4669 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361620 4669 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361628 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52twn\" (UniqueName: \"kubernetes.io/projected/4933ae0b-dda6-44e1-a551-730e803bb7d5-kube-api-access-52twn\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361641 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361660 4669 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361671 4669 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4933ae0b-dda6-44e1-a551-730e803bb7d5-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361682 4669 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361695 4669 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361706 4669 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.361714 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4933ae0b-dda6-44e1-a551-730e803bb7d5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.654554 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" event={"ID":"4933ae0b-dda6-44e1-a551-730e803bb7d5","Type":"ContainerDied","Data":"d203367db94e233f653183cc6f027da4441d539749984e0e521fb3fa264aee9e"} Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.654678 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d203367db94e233f653183cc6f027da4441d539749984e0e521fb3fa264aee9e" Oct 10 10:05:48 crc kubenswrapper[4669]: I1010 10:05:48.654687 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.762671 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 10 10:06:03 crc kubenswrapper[4669]: E1010 10:06:03.763501 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4933ae0b-dda6-44e1-a551-730e803bb7d5" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.763514 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="4933ae0b-dda6-44e1-a551-730e803bb7d5" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 10 10:06:03 crc kubenswrapper[4669]: E1010 10:06:03.763528 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="extract-content" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.763533 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="extract-content" Oct 10 10:06:03 crc kubenswrapper[4669]: E1010 10:06:03.763566 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="extract-utilities" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.763572 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="extract-utilities" Oct 10 10:06:03 crc kubenswrapper[4669]: E1010 10:06:03.763600 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="registry-server" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.763606 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="registry-server" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.763769 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="4933ae0b-dda6-44e1-a551-730e803bb7d5" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.763785 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3d6264-2366-49fa-9564-70306d79c5c8" containerName="registry-server" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.764778 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.767576 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.767737 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.769287 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.770704 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.771921 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.828796 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.828841 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.873707 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.874328 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.874430 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.874516 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-scripts\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.874620 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.874730 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.874823 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.874986 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-run\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875058 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c642048f-8f7f-4355-bd06-e862839e25e9-ceph\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875121 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5m2p\" (UniqueName: \"kubernetes.io/projected/3c92ae00-f31f-4d27-aaca-8d3920a345eb-kube-api-access-p5m2p\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875153 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t7vc\" (UniqueName: \"kubernetes.io/projected/c642048f-8f7f-4355-bd06-e862839e25e9-kube-api-access-9t7vc\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875218 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875256 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875294 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875353 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875375 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875397 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875423 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875472 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875498 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3c92ae00-f31f-4d27-aaca-8d3920a345eb-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875527 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-run\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875550 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-config-data\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875685 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875732 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875755 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875778 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875796 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875819 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-sys\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875857 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-dev\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875909 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-lib-modules\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875930 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.875972 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977569 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977654 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3c92ae00-f31f-4d27-aaca-8d3920a345eb-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977673 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-run\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977691 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-config-data\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977715 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977739 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977759 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977774 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977792 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977811 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-sys\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977826 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-dev\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977832 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-run\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.978053 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.978061 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.977855 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-lib-modules\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.978651 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.979259 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.979459 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.979597 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.979705 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.979799 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-scripts\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.979914 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.980482 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.980669 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.980792 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-run\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.980906 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c642048f-8f7f-4355-bd06-e862839e25e9-ceph\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981055 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5m2p\" (UniqueName: \"kubernetes.io/projected/3c92ae00-f31f-4d27-aaca-8d3920a345eb-kube-api-access-p5m2p\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981160 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t7vc\" (UniqueName: \"kubernetes.io/projected/c642048f-8f7f-4355-bd06-e862839e25e9-kube-api-access-9t7vc\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981287 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981420 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981542 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981714 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981826 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.981951 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.982059 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.982849 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.982913 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-run\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.980033 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.980073 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-lib-modules\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.979377 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-sys\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.983967 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.983999 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3c92ae00-f31f-4d27-aaca-8d3920a345eb-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.984272 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.984623 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.986136 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.986163 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.986550 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.986606 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.986624 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.987643 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c642048f-8f7f-4355-bd06-e862839e25e9-ceph\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:03 crc kubenswrapper[4669]: I1010 10:06:03.991294 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992203 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-config-data\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992262 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992332 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992347 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992370 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992382 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992393 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c642048f-8f7f-4355-bd06-e862839e25e9-dev\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.992393 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3c92ae00-f31f-4d27-aaca-8d3920a345eb-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.995559 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.998104 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c92ae00-f31f-4d27-aaca-8d3920a345eb-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:03.999201 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c642048f-8f7f-4355-bd06-e862839e25e9-scripts\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.002824 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5m2p\" (UniqueName: \"kubernetes.io/projected/3c92ae00-f31f-4d27-aaca-8d3920a345eb-kube-api-access-p5m2p\") pod \"cinder-volume-volume1-0\" (UID: \"3c92ae00-f31f-4d27-aaca-8d3920a345eb\") " pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.003163 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t7vc\" (UniqueName: \"kubernetes.io/projected/c642048f-8f7f-4355-bd06-e862839e25e9-kube-api-access-9t7vc\") pod \"cinder-backup-0\" (UID: \"c642048f-8f7f-4355-bd06-e862839e25e9\") " pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.083212 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.101681 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.643967 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.653014 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.657802 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4lx7j" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.657818 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.658631 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.658801 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.660296 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-mbmzv"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.661541 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mbmzv" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.713063 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.771836 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.786971 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.788120 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mbmzv"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.792223 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.792402 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.825781 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.825884 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.825903 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2942\" (UniqueName: \"kubernetes.io/projected/bbdaa724-fc17-4ef9-9c22-5949776c81f4-kube-api-access-t2942\") pod \"manila-db-create-mbmzv\" (UID: \"bbdaa724-fc17-4ef9-9c22-5949776c81f4\") " pod="openstack/manila-db-create-mbmzv" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.825925 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsbzg\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-kube-api-access-xsbzg\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.825948 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.825978 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.826012 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-ceph\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.826032 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.826051 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-logs\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.826084 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.858172 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.878535 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:04 crc kubenswrapper[4669]: E1010 10:06:04.879707 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-xsbzg logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="8ab6b1ae-0833-4ae5-937f-1d877cdfee62" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.909176 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bcdd76bc7-s8ksq"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.911256 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.919218 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.919545 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.919730 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-x7ww8" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.920620 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bcdd76bc7-s8ksq"] Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927546 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927675 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkwq6\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-kube-api-access-kkwq6\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927738 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927787 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927811 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-ceph\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927852 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927882 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-logs\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927936 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927966 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.928087 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.928549 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.928761 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-logs\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.927990 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.941660 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.941738 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.943718 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.943962 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.944029 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.944321 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.944375 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2942\" (UniqueName: \"kubernetes.io/projected/bbdaa724-fc17-4ef9-9c22-5949776c81f4-kube-api-access-t2942\") pod \"manila-db-create-mbmzv\" (UID: \"bbdaa724-fc17-4ef9-9c22-5949776c81f4\") " pod="openstack/manila-db-create-mbmzv" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.944406 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.944434 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsbzg\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-kube-api-access-xsbzg\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.944474 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-logs\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.950613 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.950719 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-ceph\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.958151 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.963441 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.973878 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.977336 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:04 crc kubenswrapper[4669]: E1010 10:06:04.978390 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-kkwq6 logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="ecf3158d-1732-4fb7-87cd-80b68c2b6583" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.996649 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2942\" (UniqueName: \"kubernetes.io/projected/bbdaa724-fc17-4ef9-9c22-5949776c81f4-kube-api-access-t2942\") pod \"manila-db-create-mbmzv\" (UID: \"bbdaa724-fc17-4ef9-9c22-5949776c81f4\") " pod="openstack/manila-db-create-mbmzv" Oct 10 10:06:04 crc kubenswrapper[4669]: I1010 10:06:04.997572 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsbzg\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-kube-api-access-xsbzg\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.035279 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-56bb6b7c6f-x4hxx"] Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.037554 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.037877 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mbmzv" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057432 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057478 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-config-data\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057507 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-logs\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057558 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkwq6\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-kube-api-access-kkwq6\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057595 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-scripts\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057654 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057698 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-horizon-secret-key\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057722 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057745 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057772 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfnl8\" (UniqueName: \"kubernetes.io/projected/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-kube-api-access-dfnl8\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057792 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057825 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-logs\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057877 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.057909 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.058083 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-logs\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.059958 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.061076 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.070573 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.070830 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.078286 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.078888 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.086312 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkwq6\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-kube-api-access-kkwq6\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.110455 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.127503 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.134180 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56bb6b7c6f-x4hxx"] Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.144950 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.160838 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-config-data\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.160907 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-scripts\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.160966 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-config-data\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.161002 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrqrx\" (UniqueName: \"kubernetes.io/projected/d8cd5531-5b53-4fe7-a052-e0229353184d-kube-api-access-qrqrx\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.161029 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-scripts\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.161078 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8cd5531-5b53-4fe7-a052-e0229353184d-logs\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.161100 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-horizon-secret-key\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.161129 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfnl8\" (UniqueName: \"kubernetes.io/projected/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-kube-api-access-dfnl8\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.161158 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-logs\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.161187 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8cd5531-5b53-4fe7-a052-e0229353184d-horizon-secret-key\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.162353 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-config-data\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.163094 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-scripts\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.163308 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-logs\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.164494 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.174539 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.176137 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-horizon-secret-key\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.191248 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfnl8\" (UniqueName: \"kubernetes.io/projected/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-kube-api-access-dfnl8\") pod \"horizon-5bcdd76bc7-s8ksq\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.195308 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.263185 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8cd5531-5b53-4fe7-a052-e0229353184d-logs\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.263282 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8cd5531-5b53-4fe7-a052-e0229353184d-horizon-secret-key\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.263321 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-config-data\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.263338 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-scripts\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.263398 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrqrx\" (UniqueName: \"kubernetes.io/projected/d8cd5531-5b53-4fe7-a052-e0229353184d-kube-api-access-qrqrx\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.264066 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8cd5531-5b53-4fe7-a052-e0229353184d-logs\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.265479 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-config-data\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.266034 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-scripts\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.272009 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8cd5531-5b53-4fe7-a052-e0229353184d-horizon-secret-key\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.290434 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrqrx\" (UniqueName: \"kubernetes.io/projected/d8cd5531-5b53-4fe7-a052-e0229353184d-kube-api-access-qrqrx\") pod \"horizon-56bb6b7c6f-x4hxx\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.356279 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.550526 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:05 crc kubenswrapper[4669]: I1010 10:06:05.710883 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mbmzv"] Oct 10 10:06:05 crc kubenswrapper[4669]: W1010 10:06:05.736735 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbdaa724_fc17_4ef9_9c22_5949776c81f4.slice/crio-3d7cfc2cd74b4d99a9ca80289076422cf2ac49dca1ac0b9aba1dab89d5d558e3 WatchSource:0}: Error finding container 3d7cfc2cd74b4d99a9ca80289076422cf2ac49dca1ac0b9aba1dab89d5d558e3: Status 404 returned error can't find the container with id 3d7cfc2cd74b4d99a9ca80289076422cf2ac49dca1ac0b9aba1dab89d5d558e3 Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.916455 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mbmzv" event={"ID":"bbdaa724-fc17-4ef9-9c22-5949776c81f4","Type":"ContainerStarted","Data":"3d7cfc2cd74b4d99a9ca80289076422cf2ac49dca1ac0b9aba1dab89d5d558e3"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.926863 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c642048f-8f7f-4355-bd06-e862839e25e9","Type":"ContainerStarted","Data":"b6ec7f3ca6548eb989c956e0e1a38fa4367a512c2f5210e64bf897190a18e5f4"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.935870 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.936074 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3c92ae00-f31f-4d27-aaca-8d3920a345eb","Type":"ContainerStarted","Data":"b97947891dfee1d3081707569f1d1b2f667e6d8dde71ec9ec60f6b6ebd73cea8"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.936343 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.940703 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bcdd76bc7-s8ksq"] Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.956416 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.967677 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.995945 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-logs\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.995995 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-httpd-run\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996022 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996078 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsbzg\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-kube-api-access-xsbzg\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996248 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-ceph\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996325 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-public-tls-certs\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996365 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-scripts\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996388 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-combined-ca-bundle\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996436 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-config-data\") pod \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\" (UID: \"8ab6b1ae-0833-4ae5-937f-1d877cdfee62\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.996496 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:05.997067 4669 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.000144 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-logs" (OuterVolumeSpecName: "logs") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.002375 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-kube-api-access-xsbzg" (OuterVolumeSpecName: "kube-api-access-xsbzg") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "kube-api-access-xsbzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.003161 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-config-data" (OuterVolumeSpecName: "config-data") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.003210 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.003365 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.003678 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-scripts" (OuterVolumeSpecName: "scripts") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.004178 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.020959 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-ceph" (OuterVolumeSpecName: "ceph") pod "8ab6b1ae-0833-4ae5-937f-1d877cdfee62" (UID: "8ab6b1ae-0833-4ae5-937f-1d877cdfee62"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098245 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098314 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-ceph\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098417 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-combined-ca-bundle\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098483 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-scripts\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098525 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-config-data\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098664 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-internal-tls-certs\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098704 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkwq6\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-kube-api-access-kkwq6\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098739 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-httpd-run\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.098863 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-logs\") pod \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\" (UID: \"ecf3158d-1732-4fb7-87cd-80b68c2b6583\") " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100026 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100045 4669 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100055 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100063 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100074 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100082 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-logs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100100 4669 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.100111 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsbzg\" (UniqueName: \"kubernetes.io/projected/8ab6b1ae-0833-4ae5-937f-1d877cdfee62-kube-api-access-xsbzg\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.101945 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.102496 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-logs" (OuterVolumeSpecName: "logs") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.107900 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-kube-api-access-kkwq6" (OuterVolumeSpecName: "kube-api-access-kkwq6") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "kube-api-access-kkwq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.109076 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-scripts" (OuterVolumeSpecName: "scripts") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.109134 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-config-data" (OuterVolumeSpecName: "config-data") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.110550 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.112572 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-ceph" (OuterVolumeSpecName: "ceph") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.113895 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.116496 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ecf3158d-1732-4fb7-87cd-80b68c2b6583" (UID: "ecf3158d-1732-4fb7-87cd-80b68c2b6583"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.133796 4669 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.195164 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56bb6b7c6f-x4hxx"] Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202088 4669 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202120 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkwq6\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-kube-api-access-kkwq6\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202133 4669 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202141 4669 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202149 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecf3158d-1732-4fb7-87cd-80b68c2b6583-logs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202184 4669 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202193 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ecf3158d-1732-4fb7-87cd-80b68c2b6583-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202201 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202209 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.202219 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf3158d-1732-4fb7-87cd-80b68c2b6583-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.240239 4669 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.307843 4669 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:06 crc kubenswrapper[4669]: W1010 10:06:06.403595 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8cd5531_5b53_4fe7_a052_e0229353184d.slice/crio-faec2d9f0ce6b72e3fcd1ee6866963bd6b371bf0c5f129f28638f39bcbf7e652 WatchSource:0}: Error finding container faec2d9f0ce6b72e3fcd1ee6866963bd6b371bf0c5f129f28638f39bcbf7e652: Status 404 returned error can't find the container with id faec2d9f0ce6b72e3fcd1ee6866963bd6b371bf0c5f129f28638f39bcbf7e652 Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.951744 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56bb6b7c6f-x4hxx" event={"ID":"d8cd5531-5b53-4fe7-a052-e0229353184d","Type":"ContainerStarted","Data":"faec2d9f0ce6b72e3fcd1ee6866963bd6b371bf0c5f129f28638f39bcbf7e652"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.955918 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bcdd76bc7-s8ksq" event={"ID":"3d1deb67-bb78-4a89-9561-4e916e5e9ad4","Type":"ContainerStarted","Data":"a6e76c17a369e91e49cd820495e125ccc8a31acae5e1ef71e20817c4e0f01abf"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.961193 4669 generic.go:334] "Generic (PLEG): container finished" podID="bbdaa724-fc17-4ef9-9c22-5949776c81f4" containerID="4204cf0fbb1bcc7c38f8bf4735936415aaae00b388b794a6f99d487922c8131d" exitCode=0 Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.961266 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mbmzv" event={"ID":"bbdaa724-fc17-4ef9-9c22-5949776c81f4","Type":"ContainerDied","Data":"4204cf0fbb1bcc7c38f8bf4735936415aaae00b388b794a6f99d487922c8131d"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.970567 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c642048f-8f7f-4355-bd06-e862839e25e9","Type":"ContainerStarted","Data":"8f3e779518198688c347e86f8882efdbd0d0bd6786e3571317ccfc370e12878d"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.970635 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c642048f-8f7f-4355-bd06-e862839e25e9","Type":"ContainerStarted","Data":"f4da83d1cae16034927cf7ac3560c6f9895699db1ae8b5922bc0a2401b223a12"} Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.992450 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.997693 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:06 crc kubenswrapper[4669]: I1010 10:06:06.997830 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3c92ae00-f31f-4d27-aaca-8d3920a345eb","Type":"ContainerStarted","Data":"8c63c077a3b2826b3057128085463733419f9792c9d139510111ad34ae2ee3aa"} Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.047048 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56bb6b7c6f-x4hxx"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.049483 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.717000235 podStartE2EDuration="4.049467759s" podCreationTimestamp="2025-10-10 10:06:03 +0000 UTC" firstStartedPulling="2025-10-10 10:06:05.110222067 +0000 UTC m=+3308.126240809" lastFinishedPulling="2025-10-10 10:06:06.442689591 +0000 UTC m=+3309.458708333" observedRunningTime="2025-10-10 10:06:07.031855504 +0000 UTC m=+3310.047874316" watchObservedRunningTime="2025-10-10 10:06:07.049467759 +0000 UTC m=+3310.065486501" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.128169 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7944cbccd6-fnqbk"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.130113 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.135918 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.138446 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7944cbccd6-fnqbk"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.183732 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.197513 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.233737 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-secret-key\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.233782 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0a1408-8b73-4070-b491-74d331bcf567-logs\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.233829 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-scripts\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.233857 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-tls-certs\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.233876 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6wwp\" (UniqueName: \"kubernetes.io/projected/af0a1408-8b73-4070-b491-74d331bcf567-kube-api-access-h6wwp\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.233891 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-config-data\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.233937 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-combined-ca-bundle\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.254782 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bcdd76bc7-s8ksq"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.336401 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6wwp\" (UniqueName: \"kubernetes.io/projected/af0a1408-8b73-4070-b491-74d331bcf567-kube-api-access-h6wwp\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.336449 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-config-data\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.336504 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-combined-ca-bundle\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.336597 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-secret-key\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.336621 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0a1408-8b73-4070-b491-74d331bcf567-logs\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.336665 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-scripts\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.336704 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-tls-certs\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.337115 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0a1408-8b73-4070-b491-74d331bcf567-logs\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.339794 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-scripts\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.340578 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.342130 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.346678 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4lx7j" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.346956 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.347084 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.347177 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.352381 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-config-data\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.359183 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-secret-key\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.359544 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-tls-certs\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.360028 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-combined-ca-bundle\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.371237 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6wwp\" (UniqueName: \"kubernetes.io/projected/af0a1408-8b73-4070-b491-74d331bcf567-kube-api-access-h6wwp\") pod \"horizon-7944cbccd6-fnqbk\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.387096 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.433513 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-566f97fbdd-plcxg"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.435780 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439550 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439671 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phd6z\" (UniqueName: \"kubernetes.io/projected/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-kube-api-access-phd6z\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439716 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439745 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439788 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-logs\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439808 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439834 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439866 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.439893 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.480362 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.491019 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.500112 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-566f97fbdd-plcxg"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.508570 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.523785 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.525330 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.527558 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.527808 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541659 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541697 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541772 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phd6z\" (UniqueName: \"kubernetes.io/projected/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-kube-api-access-phd6z\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541809 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541837 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541873 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-logs\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541891 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541915 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.541945 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.542654 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.543139 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-logs\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.555357 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.555882 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.557493 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.590618 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.592324 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.594479 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.597648 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phd6z\" (UniqueName: \"kubernetes.io/projected/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-kube-api-access-phd6z\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.601459 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/731fc6e9-0f93-4e9f-bbc0-767b6fada47b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.684134 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0040911a-8dff-49e4-99af-aafb84f7639a-scripts\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.684185 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-combined-ca-bundle\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.684226 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.684301 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.684374 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60e2fea4-35a0-45d6-bd36-e7172b6822fe-ceph\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.684422 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e2fea4-35a0-45d6-bd36-e7172b6822fe-logs\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.684447 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0040911a-8dff-49e4-99af-aafb84f7639a-config-data\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.691216 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-horizon-tls-certs\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.691266 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjth\" (UniqueName: \"kubernetes.io/projected/0040911a-8dff-49e4-99af-aafb84f7639a-kube-api-access-vkjth\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.691338 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.691416 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-horizon-secret-key\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.699712 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60e2fea4-35a0-45d6-bd36-e7172b6822fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.699780 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.699870 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n78jh\" (UniqueName: \"kubernetes.io/projected/60e2fea4-35a0-45d6-bd36-e7172b6822fe-kube-api-access-n78jh\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.699893 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.699952 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0040911a-8dff-49e4-99af-aafb84f7639a-logs\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.719720 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"731fc6e9-0f93-4e9f-bbc0-767b6fada47b\") " pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.785382 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801632 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0040911a-8dff-49e4-99af-aafb84f7639a-scripts\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801667 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-combined-ca-bundle\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801692 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801722 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801753 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60e2fea4-35a0-45d6-bd36-e7172b6822fe-ceph\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801778 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e2fea4-35a0-45d6-bd36-e7172b6822fe-logs\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801795 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0040911a-8dff-49e4-99af-aafb84f7639a-config-data\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801816 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-horizon-tls-certs\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801838 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjth\" (UniqueName: \"kubernetes.io/projected/0040911a-8dff-49e4-99af-aafb84f7639a-kube-api-access-vkjth\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801867 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801907 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-horizon-secret-key\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801923 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60e2fea4-35a0-45d6-bd36-e7172b6822fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801938 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801972 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n78jh\" (UniqueName: \"kubernetes.io/projected/60e2fea4-35a0-45d6-bd36-e7172b6822fe-kube-api-access-n78jh\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.801992 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.802027 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0040911a-8dff-49e4-99af-aafb84f7639a-logs\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.802785 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0040911a-8dff-49e4-99af-aafb84f7639a-logs\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.803370 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0040911a-8dff-49e4-99af-aafb84f7639a-scripts\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.807325 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60e2fea4-35a0-45d6-bd36-e7172b6822fe-logs\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.808305 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0040911a-8dff-49e4-99af-aafb84f7639a-config-data\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.811258 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.811379 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60e2fea4-35a0-45d6-bd36-e7172b6822fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.815713 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.816134 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60e2fea4-35a0-45d6-bd36-e7172b6822fe-ceph\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.816144 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.816668 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-horizon-tls-certs\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.817103 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.818233 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-combined-ca-bundle\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.821907 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60e2fea4-35a0-45d6-bd36-e7172b6822fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.826049 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0040911a-8dff-49e4-99af-aafb84f7639a-horizon-secret-key\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.844112 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjth\" (UniqueName: \"kubernetes.io/projected/0040911a-8dff-49e4-99af-aafb84f7639a-kube-api-access-vkjth\") pod \"horizon-566f97fbdd-plcxg\" (UID: \"0040911a-8dff-49e4-99af-aafb84f7639a\") " pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.873755 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.888211 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ab6b1ae-0833-4ae5-937f-1d877cdfee62" path="/var/lib/kubelet/pods/8ab6b1ae-0833-4ae5-937f-1d877cdfee62/volumes" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.888718 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecf3158d-1732-4fb7-87cd-80b68c2b6583" path="/var/lib/kubelet/pods/ecf3158d-1732-4fb7-87cd-80b68c2b6583/volumes" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.893051 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n78jh\" (UniqueName: \"kubernetes.io/projected/60e2fea4-35a0-45d6-bd36-e7172b6822fe-kube-api-access-n78jh\") pod \"glance-default-external-api-0\" (UID: \"60e2fea4-35a0-45d6-bd36-e7172b6822fe\") " pod="openstack/glance-default-external-api-0" Oct 10 10:06:07 crc kubenswrapper[4669]: I1010 10:06:07.957162 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.040756 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3c92ae00-f31f-4d27-aaca-8d3920a345eb","Type":"ContainerStarted","Data":"e4698328e5a7ad7347e522a3f6d4c3bd43129b75c327c00c5979f2a2dd556122"} Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.088684 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.785312656 podStartE2EDuration="5.088667329s" podCreationTimestamp="2025-10-10 10:06:03 +0000 UTC" firstStartedPulling="2025-10-10 10:06:05.145709903 +0000 UTC m=+3308.161728645" lastFinishedPulling="2025-10-10 10:06:06.449064576 +0000 UTC m=+3309.465083318" observedRunningTime="2025-10-10 10:06:08.088596607 +0000 UTC m=+3311.104615349" watchObservedRunningTime="2025-10-10 10:06:08.088667329 +0000 UTC m=+3311.104686061" Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.106219 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.180852 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7944cbccd6-fnqbk"] Oct 10 10:06:08 crc kubenswrapper[4669]: W1010 10:06:08.202851 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf0a1408_8b73_4070_b491_74d331bcf567.slice/crio-60d4dd2fdafab6673200c8a06a60c01c66092828b1193d8c2970ede1aeae985d WatchSource:0}: Error finding container 60d4dd2fdafab6673200c8a06a60c01c66092828b1193d8c2970ede1aeae985d: Status 404 returned error can't find the container with id 60d4dd2fdafab6673200c8a06a60c01c66092828b1193d8c2970ede1aeae985d Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.565817 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.620308 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mbmzv" Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.773532 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2942\" (UniqueName: \"kubernetes.io/projected/bbdaa724-fc17-4ef9-9c22-5949776c81f4-kube-api-access-t2942\") pod \"bbdaa724-fc17-4ef9-9c22-5949776c81f4\" (UID: \"bbdaa724-fc17-4ef9-9c22-5949776c81f4\") " Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.787176 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbdaa724-fc17-4ef9-9c22-5949776c81f4-kube-api-access-t2942" (OuterVolumeSpecName: "kube-api-access-t2942") pod "bbdaa724-fc17-4ef9-9c22-5949776c81f4" (UID: "bbdaa724-fc17-4ef9-9c22-5949776c81f4"). InnerVolumeSpecName "kube-api-access-t2942". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.839741 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 10:06:08 crc kubenswrapper[4669]: W1010 10:06:08.860521 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60e2fea4_35a0_45d6_bd36_e7172b6822fe.slice/crio-05b8b5cbe6a229e72fccea92bff15ab78038cdf76d117b5a2ceaa097819cb7e2 WatchSource:0}: Error finding container 05b8b5cbe6a229e72fccea92bff15ab78038cdf76d117b5a2ceaa097819cb7e2: Status 404 returned error can't find the container with id 05b8b5cbe6a229e72fccea92bff15ab78038cdf76d117b5a2ceaa097819cb7e2 Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.875353 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2942\" (UniqueName: \"kubernetes.io/projected/bbdaa724-fc17-4ef9-9c22-5949776c81f4-kube-api-access-t2942\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:08 crc kubenswrapper[4669]: I1010 10:06:08.893670 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-566f97fbdd-plcxg"] Oct 10 10:06:08 crc kubenswrapper[4669]: W1010 10:06:08.944652 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0040911a_8dff_49e4_99af_aafb84f7639a.slice/crio-732f582d8fdef89c56b19f842878663ab162888ce4e4bcf4dbd988e7588f6623 WatchSource:0}: Error finding container 732f582d8fdef89c56b19f842878663ab162888ce4e4bcf4dbd988e7588f6623: Status 404 returned error can't find the container with id 732f582d8fdef89c56b19f842878663ab162888ce4e4bcf4dbd988e7588f6623 Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.054641 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7944cbccd6-fnqbk" event={"ID":"af0a1408-8b73-4070-b491-74d331bcf567","Type":"ContainerStarted","Data":"60d4dd2fdafab6673200c8a06a60c01c66092828b1193d8c2970ede1aeae985d"} Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.062087 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mbmzv" event={"ID":"bbdaa724-fc17-4ef9-9c22-5949776c81f4","Type":"ContainerDied","Data":"3d7cfc2cd74b4d99a9ca80289076422cf2ac49dca1ac0b9aba1dab89d5d558e3"} Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.062126 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d7cfc2cd74b4d99a9ca80289076422cf2ac49dca1ac0b9aba1dab89d5d558e3" Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.062103 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mbmzv" Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.063652 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"731fc6e9-0f93-4e9f-bbc0-767b6fada47b","Type":"ContainerStarted","Data":"579b0182bbd148ba35973aa0dc3e48c2c307f878ee8e19e64963c8e3421c5c1c"} Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.066209 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60e2fea4-35a0-45d6-bd36-e7172b6822fe","Type":"ContainerStarted","Data":"05b8b5cbe6a229e72fccea92bff15ab78038cdf76d117b5a2ceaa097819cb7e2"} Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.075274 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566f97fbdd-plcxg" event={"ID":"0040911a-8dff-49e4-99af-aafb84f7639a","Type":"ContainerStarted","Data":"732f582d8fdef89c56b19f842878663ab162888ce4e4bcf4dbd988e7588f6623"} Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.084783 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 10 10:06:09 crc kubenswrapper[4669]: I1010 10:06:09.102664 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:10 crc kubenswrapper[4669]: I1010 10:06:10.107877 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"731fc6e9-0f93-4e9f-bbc0-767b6fada47b","Type":"ContainerStarted","Data":"ed94fcd02a66c716dd4b64711b6158b2385e11189104914fe3e59f36f49bae02"} Oct 10 10:06:11 crc kubenswrapper[4669]: I1010 10:06:11.118982 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60e2fea4-35a0-45d6-bd36-e7172b6822fe","Type":"ContainerStarted","Data":"7791a71aa7a5d67b3e4a2c7cf2fc3eae712bc6f8accdcfd28268fb0d9de14b59"} Oct 10 10:06:12 crc kubenswrapper[4669]: I1010 10:06:12.136341 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"731fc6e9-0f93-4e9f-bbc0-767b6fada47b","Type":"ContainerStarted","Data":"d8c065734f5e004d9e1ca63624c6cbed26afbbaab3e840363f2d9962b1c2956d"} Oct 10 10:06:12 crc kubenswrapper[4669]: I1010 10:06:12.138701 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60e2fea4-35a0-45d6-bd36-e7172b6822fe","Type":"ContainerStarted","Data":"dab467ff093d8bcb426d02e5e5b55ae2d7e17e0e72810dae49c7c743dd560b84"} Oct 10 10:06:12 crc kubenswrapper[4669]: I1010 10:06:12.179154 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.179136034 podStartE2EDuration="5.179136034s" podCreationTimestamp="2025-10-10 10:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:06:12.175783777 +0000 UTC m=+3315.191802519" watchObservedRunningTime="2025-10-10 10:06:12.179136034 +0000 UTC m=+3315.195154776" Oct 10 10:06:12 crc kubenswrapper[4669]: I1010 10:06:12.212802 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.212777812 podStartE2EDuration="5.212777812s" podCreationTimestamp="2025-10-10 10:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:06:12.208956639 +0000 UTC m=+3315.224975381" watchObservedRunningTime="2025-10-10 10:06:12.212777812 +0000 UTC m=+3315.228796554" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.372830 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.381739 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.724397 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-1e19-account-create-gllwp"] Oct 10 10:06:14 crc kubenswrapper[4669]: E1010 10:06:14.731284 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbdaa724-fc17-4ef9-9c22-5949776c81f4" containerName="mariadb-database-create" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.731315 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbdaa724-fc17-4ef9-9c22-5949776c81f4" containerName="mariadb-database-create" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.731539 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbdaa724-fc17-4ef9-9c22-5949776c81f4" containerName="mariadb-database-create" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.732236 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1e19-account-create-gllwp" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.751138 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.759606 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-1e19-account-create-gllwp"] Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.864956 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgjfs\" (UniqueName: \"kubernetes.io/projected/7fddf3cf-5756-477f-9dd1-c653f747776d-kube-api-access-rgjfs\") pod \"manila-1e19-account-create-gllwp\" (UID: \"7fddf3cf-5756-477f-9dd1-c653f747776d\") " pod="openstack/manila-1e19-account-create-gllwp" Oct 10 10:06:14 crc kubenswrapper[4669]: I1010 10:06:14.969903 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgjfs\" (UniqueName: \"kubernetes.io/projected/7fddf3cf-5756-477f-9dd1-c653f747776d-kube-api-access-rgjfs\") pod \"manila-1e19-account-create-gllwp\" (UID: \"7fddf3cf-5756-477f-9dd1-c653f747776d\") " pod="openstack/manila-1e19-account-create-gllwp" Oct 10 10:06:15 crc kubenswrapper[4669]: I1010 10:06:14.998376 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgjfs\" (UniqueName: \"kubernetes.io/projected/7fddf3cf-5756-477f-9dd1-c653f747776d-kube-api-access-rgjfs\") pod \"manila-1e19-account-create-gllwp\" (UID: \"7fddf3cf-5756-477f-9dd1-c653f747776d\") " pod="openstack/manila-1e19-account-create-gllwp" Oct 10 10:06:15 crc kubenswrapper[4669]: I1010 10:06:15.066472 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1e19-account-create-gllwp" Oct 10 10:06:17 crc kubenswrapper[4669]: I1010 10:06:17.787038 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:17 crc kubenswrapper[4669]: I1010 10:06:17.787682 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:17 crc kubenswrapper[4669]: I1010 10:06:17.841835 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:17 crc kubenswrapper[4669]: I1010 10:06:17.842354 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:17 crc kubenswrapper[4669]: I1010 10:06:17.958827 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 10:06:17 crc kubenswrapper[4669]: I1010 10:06:17.958925 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 10:06:17 crc kubenswrapper[4669]: I1010 10:06:17.997446 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 10:06:18 crc kubenswrapper[4669]: I1010 10:06:18.009927 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 10:06:18 crc kubenswrapper[4669]: I1010 10:06:18.190947 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:18 crc kubenswrapper[4669]: I1010 10:06:18.191000 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 10:06:18 crc kubenswrapper[4669]: I1010 10:06:18.191014 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:18 crc kubenswrapper[4669]: I1010 10:06:18.191028 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 10:06:20 crc kubenswrapper[4669]: I1010 10:06:20.630888 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-1e19-account-create-gllwp"] Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.289702 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566f97fbdd-plcxg" event={"ID":"0040911a-8dff-49e4-99af-aafb84f7639a","Type":"ContainerStarted","Data":"b73d78e75c8cd4344dff1b451682fdf753cf6ead2b12de9936b9827e648d5e68"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.290023 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-566f97fbdd-plcxg" event={"ID":"0040911a-8dff-49e4-99af-aafb84f7639a","Type":"ContainerStarted","Data":"e44ad6c5863f4b34670c73baa4a80b07f2409a825addbe066b279563c3592ee9"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.293388 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56bb6b7c6f-x4hxx" event={"ID":"d8cd5531-5b53-4fe7-a052-e0229353184d","Type":"ContainerStarted","Data":"9766feaf40e48542a096684c3ca9b44d3bea50756217264471b768047903fac3"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.293439 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56bb6b7c6f-x4hxx" event={"ID":"d8cd5531-5b53-4fe7-a052-e0229353184d","Type":"ContainerStarted","Data":"a6d58339f9ad409f7dddb3adbbb7f6ffa8664790c3ed4adb5619217c62271fac"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.293662 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-56bb6b7c6f-x4hxx" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon-log" containerID="cri-o://a6d58339f9ad409f7dddb3adbbb7f6ffa8664790c3ed4adb5619217c62271fac" gracePeriod=30 Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.293994 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-56bb6b7c6f-x4hxx" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon" containerID="cri-o://9766feaf40e48542a096684c3ca9b44d3bea50756217264471b768047903fac3" gracePeriod=30 Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.305395 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bcdd76bc7-s8ksq" event={"ID":"3d1deb67-bb78-4a89-9561-4e916e5e9ad4","Type":"ContainerStarted","Data":"8bd1daf033e5f0a1152d540c79ded0c7e56f4f58de2643d12a7501818072dbfe"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.305439 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bcdd76bc7-s8ksq" event={"ID":"3d1deb67-bb78-4a89-9561-4e916e5e9ad4","Type":"ContainerStarted","Data":"50243f1d5bad56559ea2a44eb8837f9acc033af9e4ff55ea3d9a81249f8fb61a"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.305500 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bcdd76bc7-s8ksq" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon-log" containerID="cri-o://50243f1d5bad56559ea2a44eb8837f9acc033af9e4ff55ea3d9a81249f8fb61a" gracePeriod=30 Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.305533 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bcdd76bc7-s8ksq" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon" containerID="cri-o://8bd1daf033e5f0a1152d540c79ded0c7e56f4f58de2643d12a7501818072dbfe" gracePeriod=30 Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.311527 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7944cbccd6-fnqbk" event={"ID":"af0a1408-8b73-4070-b491-74d331bcf567","Type":"ContainerStarted","Data":"c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.311569 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7944cbccd6-fnqbk" event={"ID":"af0a1408-8b73-4070-b491-74d331bcf567","Type":"ContainerStarted","Data":"5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.315322 4669 generic.go:334] "Generic (PLEG): container finished" podID="7fddf3cf-5756-477f-9dd1-c653f747776d" containerID="872f3eef90fb9aa1b9705d34e1141ed1e6d6830ba4b7ca0f74a70c61e5838f55" exitCode=0 Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.315366 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1e19-account-create-gllwp" event={"ID":"7fddf3cf-5756-477f-9dd1-c653f747776d","Type":"ContainerDied","Data":"872f3eef90fb9aa1b9705d34e1141ed1e6d6830ba4b7ca0f74a70c61e5838f55"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.315401 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1e19-account-create-gllwp" event={"ID":"7fddf3cf-5756-477f-9dd1-c653f747776d","Type":"ContainerStarted","Data":"fe28478cfa5538b9feda0c7ddeea267c6a74f1bcf55003892e0fdec78fa5e7d6"} Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.320856 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-566f97fbdd-plcxg" podStartSLOduration=3.215842372 podStartE2EDuration="14.320833663s" podCreationTimestamp="2025-10-10 10:06:07 +0000 UTC" firstStartedPulling="2025-10-10 10:06:08.950778156 +0000 UTC m=+3311.966796898" lastFinishedPulling="2025-10-10 10:06:20.055769427 +0000 UTC m=+3323.071788189" observedRunningTime="2025-10-10 10:06:21.319840411 +0000 UTC m=+3324.335859153" watchObservedRunningTime="2025-10-10 10:06:21.320833663 +0000 UTC m=+3324.336852405" Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.346765 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-56bb6b7c6f-x4hxx" podStartSLOduration=3.646799383 podStartE2EDuration="17.346747962s" podCreationTimestamp="2025-10-10 10:06:04 +0000 UTC" firstStartedPulling="2025-10-10 10:06:06.434423866 +0000 UTC m=+3309.450442598" lastFinishedPulling="2025-10-10 10:06:20.134372415 +0000 UTC m=+3323.150391177" observedRunningTime="2025-10-10 10:06:21.345955138 +0000 UTC m=+3324.361973880" watchObservedRunningTime="2025-10-10 10:06:21.346747962 +0000 UTC m=+3324.362766704" Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.374764 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5bcdd76bc7-s8ksq" podStartSLOduration=3.252187022 podStartE2EDuration="17.374747369s" podCreationTimestamp="2025-10-10 10:06:04 +0000 UTC" firstStartedPulling="2025-10-10 10:06:05.932328042 +0000 UTC m=+3308.948346784" lastFinishedPulling="2025-10-10 10:06:20.054888379 +0000 UTC m=+3323.070907131" observedRunningTime="2025-10-10 10:06:21.373062176 +0000 UTC m=+3324.389080918" watchObservedRunningTime="2025-10-10 10:06:21.374747369 +0000 UTC m=+3324.390766121" Oct 10 10:06:21 crc kubenswrapper[4669]: I1010 10:06:21.428759 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7944cbccd6-fnqbk" podStartSLOduration=2.5812594129999997 podStartE2EDuration="14.428737479s" podCreationTimestamp="2025-10-10 10:06:07 +0000 UTC" firstStartedPulling="2025-10-10 10:06:08.233297962 +0000 UTC m=+3311.249316704" lastFinishedPulling="2025-10-10 10:06:20.080776028 +0000 UTC m=+3323.096794770" observedRunningTime="2025-10-10 10:06:21.408264203 +0000 UTC m=+3324.424282945" watchObservedRunningTime="2025-10-10 10:06:21.428737479 +0000 UTC m=+3324.444756221" Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.636501 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.637142 4669 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.655670 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.731086 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1e19-account-create-gllwp" Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.882358 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.882796 4669 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.903312 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgjfs\" (UniqueName: \"kubernetes.io/projected/7fddf3cf-5756-477f-9dd1-c653f747776d-kube-api-access-rgjfs\") pod \"7fddf3cf-5756-477f-9dd1-c653f747776d\" (UID: \"7fddf3cf-5756-477f-9dd1-c653f747776d\") " Oct 10 10:06:22 crc kubenswrapper[4669]: I1010 10:06:22.916780 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fddf3cf-5756-477f-9dd1-c653f747776d-kube-api-access-rgjfs" (OuterVolumeSpecName: "kube-api-access-rgjfs") pod "7fddf3cf-5756-477f-9dd1-c653f747776d" (UID: "7fddf3cf-5756-477f-9dd1-c653f747776d"). InnerVolumeSpecName "kube-api-access-rgjfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:23 crc kubenswrapper[4669]: I1010 10:06:23.005752 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgjfs\" (UniqueName: \"kubernetes.io/projected/7fddf3cf-5756-477f-9dd1-c653f747776d-kube-api-access-rgjfs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:23 crc kubenswrapper[4669]: I1010 10:06:23.064490 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 10:06:23 crc kubenswrapper[4669]: I1010 10:06:23.332930 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1e19-account-create-gllwp" Oct 10 10:06:23 crc kubenswrapper[4669]: I1010 10:06:23.341396 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1e19-account-create-gllwp" event={"ID":"7fddf3cf-5756-477f-9dd1-c653f747776d","Type":"ContainerDied","Data":"fe28478cfa5538b9feda0c7ddeea267c6a74f1bcf55003892e0fdec78fa5e7d6"} Oct 10 10:06:23 crc kubenswrapper[4669]: I1010 10:06:23.341439 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe28478cfa5538b9feda0c7ddeea267c6a74f1bcf55003892e0fdec78fa5e7d6" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.149807 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-b47bn"] Oct 10 10:06:25 crc kubenswrapper[4669]: E1010 10:06:25.151897 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fddf3cf-5756-477f-9dd1-c653f747776d" containerName="mariadb-account-create" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.151989 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fddf3cf-5756-477f-9dd1-c653f747776d" containerName="mariadb-account-create" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.152242 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fddf3cf-5756-477f-9dd1-c653f747776d" containerName="mariadb-account-create" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.152994 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.156718 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-j92w8" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.156900 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.170451 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-b47bn"] Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.256058 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-combined-ca-bundle\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.256332 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-config-data\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.256400 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-job-config-data\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.256556 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-988xf\" (UniqueName: \"kubernetes.io/projected/f753198c-0e37-4047-915c-3511c37c70bc-kube-api-access-988xf\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.358040 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.358262 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-job-config-data\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.358342 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-988xf\" (UniqueName: \"kubernetes.io/projected/f753198c-0e37-4047-915c-3511c37c70bc-kube-api-access-988xf\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.358525 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-combined-ca-bundle\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.358573 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-config-data\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.365287 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-config-data\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.366146 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-combined-ca-bundle\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.387317 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-job-config-data\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.393087 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-988xf\" (UniqueName: \"kubernetes.io/projected/f753198c-0e37-4047-915c-3511c37c70bc-kube-api-access-988xf\") pod \"manila-db-sync-b47bn\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.470338 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:25 crc kubenswrapper[4669]: I1010 10:06:25.551030 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:26 crc kubenswrapper[4669]: I1010 10:06:26.559862 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-b47bn"] Oct 10 10:06:27 crc kubenswrapper[4669]: I1010 10:06:27.374631 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-b47bn" event={"ID":"f753198c-0e37-4047-915c-3511c37c70bc","Type":"ContainerStarted","Data":"40d672fdcef9a41c450449126c2569c52446882faed0dab7f70145facc4ace6b"} Oct 10 10:06:27 crc kubenswrapper[4669]: I1010 10:06:27.481417 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:27 crc kubenswrapper[4669]: I1010 10:06:27.481554 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:06:28 crc kubenswrapper[4669]: I1010 10:06:28.106973 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:28 crc kubenswrapper[4669]: I1010 10:06:28.108236 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:06:34 crc kubenswrapper[4669]: I1010 10:06:34.474246 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-b47bn" event={"ID":"f753198c-0e37-4047-915c-3511c37c70bc","Type":"ContainerStarted","Data":"fb5b226bd80f4b0b3e5d065a74c50163a2d3ace424db322cbf66283f32bf3f7f"} Oct 10 10:06:34 crc kubenswrapper[4669]: I1010 10:06:34.496503 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-b47bn" podStartSLOduration=2.368920819 podStartE2EDuration="9.496470854s" podCreationTimestamp="2025-10-10 10:06:25 +0000 UTC" firstStartedPulling="2025-10-10 10:06:26.572007151 +0000 UTC m=+3329.588025893" lastFinishedPulling="2025-10-10 10:06:33.699557186 +0000 UTC m=+3336.715575928" observedRunningTime="2025-10-10 10:06:34.491081431 +0000 UTC m=+3337.507100173" watchObservedRunningTime="2025-10-10 10:06:34.496470854 +0000 UTC m=+3337.512489586" Oct 10 10:06:37 crc kubenswrapper[4669]: I1010 10:06:37.482931 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7944cbccd6-fnqbk" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 10 10:06:38 crc kubenswrapper[4669]: I1010 10:06:38.109242 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-566f97fbdd-plcxg" podUID="0040911a-8dff-49e4-99af-aafb84f7639a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.249:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.249:8443: connect: connection refused" Oct 10 10:06:45 crc kubenswrapper[4669]: I1010 10:06:45.582705 4669 generic.go:334] "Generic (PLEG): container finished" podID="f753198c-0e37-4047-915c-3511c37c70bc" containerID="fb5b226bd80f4b0b3e5d065a74c50163a2d3ace424db322cbf66283f32bf3f7f" exitCode=0 Oct 10 10:06:45 crc kubenswrapper[4669]: I1010 10:06:45.582791 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-b47bn" event={"ID":"f753198c-0e37-4047-915c-3511c37c70bc","Type":"ContainerDied","Data":"fb5b226bd80f4b0b3e5d065a74c50163a2d3ace424db322cbf66283f32bf3f7f"} Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.383557 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.447178 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-combined-ca-bundle\") pod \"f753198c-0e37-4047-915c-3511c37c70bc\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.447468 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-988xf\" (UniqueName: \"kubernetes.io/projected/f753198c-0e37-4047-915c-3511c37c70bc-kube-api-access-988xf\") pod \"f753198c-0e37-4047-915c-3511c37c70bc\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.447564 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-job-config-data\") pod \"f753198c-0e37-4047-915c-3511c37c70bc\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.447602 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-config-data\") pod \"f753198c-0e37-4047-915c-3511c37c70bc\" (UID: \"f753198c-0e37-4047-915c-3511c37c70bc\") " Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.472724 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "f753198c-0e37-4047-915c-3511c37c70bc" (UID: "f753198c-0e37-4047-915c-3511c37c70bc"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.473481 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f753198c-0e37-4047-915c-3511c37c70bc-kube-api-access-988xf" (OuterVolumeSpecName: "kube-api-access-988xf") pod "f753198c-0e37-4047-915c-3511c37c70bc" (UID: "f753198c-0e37-4047-915c-3511c37c70bc"). InnerVolumeSpecName "kube-api-access-988xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.482776 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-config-data" (OuterVolumeSpecName: "config-data") pod "f753198c-0e37-4047-915c-3511c37c70bc" (UID: "f753198c-0e37-4047-915c-3511c37c70bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.482772 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7944cbccd6-fnqbk" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.491093 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f753198c-0e37-4047-915c-3511c37c70bc" (UID: "f753198c-0e37-4047-915c-3511c37c70bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.551285 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-988xf\" (UniqueName: \"kubernetes.io/projected/f753198c-0e37-4047-915c-3511c37c70bc-kube-api-access-988xf\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.551319 4669 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.551328 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.551337 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f753198c-0e37-4047-915c-3511c37c70bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.602300 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-b47bn" event={"ID":"f753198c-0e37-4047-915c-3511c37c70bc","Type":"ContainerDied","Data":"40d672fdcef9a41c450449126c2569c52446882faed0dab7f70145facc4ace6b"} Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.602353 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40d672fdcef9a41c450449126c2569c52446882faed0dab7f70145facc4ace6b" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.602358 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-b47bn" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.934576 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:06:47 crc kubenswrapper[4669]: E1010 10:06:47.953135 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f753198c-0e37-4047-915c-3511c37c70bc" containerName="manila-db-sync" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.953234 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f753198c-0e37-4047-915c-3511c37c70bc" containerName="manila-db-sync" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.953569 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f753198c-0e37-4047-915c-3511c37c70bc" containerName="manila-db-sync" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.954907 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.959185 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-j92w8" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.959575 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.959755 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 10 10:06:47 crc kubenswrapper[4669]: I1010 10:06:47.960215 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.001223 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.019860 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.022111 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.033504 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063278 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063343 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063372 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dxxt\" (UniqueName: \"kubernetes.io/projected/e5258eea-9c8f-4551-9874-af8c46837ec2-kube-api-access-5dxxt\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063441 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cdls\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-kube-api-access-5cdls\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063535 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063577 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-ceph\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063698 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063762 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063788 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063831 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-scripts\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063861 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063916 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5258eea-9c8f-4551-9874-af8c46837ec2-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063945 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.063968 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-scripts\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.083194 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179524 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179638 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179661 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179693 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-scripts\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179728 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179776 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5258eea-9c8f-4551-9874-af8c46837ec2-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179809 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179828 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-scripts\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179927 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179948 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.179968 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dxxt\" (UniqueName: \"kubernetes.io/projected/e5258eea-9c8f-4551-9874-af8c46837ec2-kube-api-access-5dxxt\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.180022 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cdls\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-kube-api-access-5cdls\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.180136 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.180164 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-ceph\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.183113 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.211455 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5258eea-9c8f-4551-9874-af8c46837ec2-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.218635 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.218837 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.219330 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.219730 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-scripts\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.240372 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.240757 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-ceph\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.241010 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.241852 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.242470 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.243069 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-scripts\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.273845 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58d5f4b747-fl954"] Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.274793 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cdls\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-kube-api-access-5cdls\") pod \"manila-share-share1-0\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.276136 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.311210 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dxxt\" (UniqueName: \"kubernetes.io/projected/e5258eea-9c8f-4551-9874-af8c46837ec2-kube-api-access-5dxxt\") pod \"manila-scheduler-0\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.313351 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58d5f4b747-fl954"] Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.340463 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.348065 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.355056 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.359269 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403257 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-openstack-edpm-ipam\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403335 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data-custom\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403407 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403449 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-config\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403473 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-ovsdbserver-sb\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403502 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkvhp\" (UniqueName: \"kubernetes.io/projected/2dff0743-e2de-497b-8b6e-8b7773e19da5-kube-api-access-pkvhp\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403554 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-ovsdbserver-nb\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403623 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-etc-machine-id\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403646 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-logs\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403674 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-dns-svc\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403700 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-scripts\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403725 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.403760 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m29v\" (UniqueName: \"kubernetes.io/projected/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-kube-api-access-4m29v\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.409536 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.509925 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510007 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-config\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510028 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-ovsdbserver-sb\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510052 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkvhp\" (UniqueName: \"kubernetes.io/projected/2dff0743-e2de-497b-8b6e-8b7773e19da5-kube-api-access-pkvhp\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510101 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-ovsdbserver-nb\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510146 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-etc-machine-id\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510179 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-logs\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510203 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-dns-svc\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510221 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-scripts\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510242 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510271 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m29v\" (UniqueName: \"kubernetes.io/projected/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-kube-api-access-4m29v\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510313 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-openstack-edpm-ipam\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.510339 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data-custom\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.511640 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-config\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.514251 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-logs\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.514913 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-ovsdbserver-sb\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.515724 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-ovsdbserver-nb\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.515759 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-etc-machine-id\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.517781 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.520041 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-dns-svc\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.521391 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/2dff0743-e2de-497b-8b6e-8b7773e19da5-openstack-edpm-ipam\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.526517 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-scripts\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.531674 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data-custom\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.552556 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.574766 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m29v\" (UniqueName: \"kubernetes.io/projected/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-kube-api-access-4m29v\") pod \"manila-api-0\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " pod="openstack/manila-api-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.581366 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.583289 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkvhp\" (UniqueName: \"kubernetes.io/projected/2dff0743-e2de-497b-8b6e-8b7773e19da5-kube-api-access-pkvhp\") pod \"dnsmasq-dns-58d5f4b747-fl954\" (UID: \"2dff0743-e2de-497b-8b6e-8b7773e19da5\") " pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.721388 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:48 crc kubenswrapper[4669]: I1010 10:06:48.840173 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 10:06:49 crc kubenswrapper[4669]: I1010 10:06:49.401323 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:06:49 crc kubenswrapper[4669]: I1010 10:06:49.447233 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:06:49 crc kubenswrapper[4669]: I1010 10:06:49.527557 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58d5f4b747-fl954"] Oct 10 10:06:49 crc kubenswrapper[4669]: I1010 10:06:49.641415 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" event={"ID":"2dff0743-e2de-497b-8b6e-8b7773e19da5","Type":"ContainerStarted","Data":"0f83b38989f89343a90c9a2af72d6f1beaa3758b6e5f10c7acea234201135279"} Oct 10 10:06:49 crc kubenswrapper[4669]: I1010 10:06:49.646570 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e5258eea-9c8f-4551-9874-af8c46837ec2","Type":"ContainerStarted","Data":"8818281a9fe7920d91feafd13536aa2458f8dc119161ba13315a60eeae155189"} Oct 10 10:06:49 crc kubenswrapper[4669]: I1010 10:06:49.658801 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"926f04a8-ae97-4fe4-aaaa-edbec3948f79","Type":"ContainerStarted","Data":"db5622e287588f7b75837593953a73674b9837a8bf3d1a9d73374f403f6b74c1"} Oct 10 10:06:49 crc kubenswrapper[4669]: I1010 10:06:49.820209 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:50 crc kubenswrapper[4669]: I1010 10:06:50.691605 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"dfb11810-489a-4b3b-9ae2-63fc4081c8fa","Type":"ContainerStarted","Data":"0f8ec6cb84fc0798faaea988751350ab1af48198e622ca958fb104f3e885a640"} Oct 10 10:06:50 crc kubenswrapper[4669]: I1010 10:06:50.698243 4669 generic.go:334] "Generic (PLEG): container finished" podID="2dff0743-e2de-497b-8b6e-8b7773e19da5" containerID="0fbbde281aa16dc8a45bc836bdbf331084b57387a389c75c44bbb1b369d49c5b" exitCode=0 Oct 10 10:06:50 crc kubenswrapper[4669]: I1010 10:06:50.698294 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" event={"ID":"2dff0743-e2de-497b-8b6e-8b7773e19da5","Type":"ContainerDied","Data":"0fbbde281aa16dc8a45bc836bdbf331084b57387a389c75c44bbb1b369d49c5b"} Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.792316 4669 generic.go:334] "Generic (PLEG): container finished" podID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerID="9766feaf40e48542a096684c3ca9b44d3bea50756217264471b768047903fac3" exitCode=137 Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.792910 4669 generic.go:334] "Generic (PLEG): container finished" podID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerID="a6d58339f9ad409f7dddb3adbbb7f6ffa8664790c3ed4adb5619217c62271fac" exitCode=137 Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.792961 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56bb6b7c6f-x4hxx" event={"ID":"d8cd5531-5b53-4fe7-a052-e0229353184d","Type":"ContainerDied","Data":"9766feaf40e48542a096684c3ca9b44d3bea50756217264471b768047903fac3"} Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.792996 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56bb6b7c6f-x4hxx" event={"ID":"d8cd5531-5b53-4fe7-a052-e0229353184d","Type":"ContainerDied","Data":"a6d58339f9ad409f7dddb3adbbb7f6ffa8664790c3ed4adb5619217c62271fac"} Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.841686 4669 generic.go:334] "Generic (PLEG): container finished" podID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerID="8bd1daf033e5f0a1152d540c79ded0c7e56f4f58de2643d12a7501818072dbfe" exitCode=137 Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.842059 4669 generic.go:334] "Generic (PLEG): container finished" podID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerID="50243f1d5bad56559ea2a44eb8837f9acc033af9e4ff55ea3d9a81249f8fb61a" exitCode=137 Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.848726 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" event={"ID":"2dff0743-e2de-497b-8b6e-8b7773e19da5","Type":"ContainerStarted","Data":"4458284ab3eceaeb04d3bc334bba527c43546af8244cbc22b40297e08c01aeaa"} Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.848792 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bcdd76bc7-s8ksq" event={"ID":"3d1deb67-bb78-4a89-9561-4e916e5e9ad4","Type":"ContainerDied","Data":"8bd1daf033e5f0a1152d540c79ded0c7e56f4f58de2643d12a7501818072dbfe"} Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.848820 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.848863 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bcdd76bc7-s8ksq" event={"ID":"3d1deb67-bb78-4a89-9561-4e916e5e9ad4","Type":"ContainerDied","Data":"50243f1d5bad56559ea2a44eb8837f9acc033af9e4ff55ea3d9a81249f8fb61a"} Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.856191 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"dfb11810-489a-4b3b-9ae2-63fc4081c8fa","Type":"ContainerStarted","Data":"d05f7c14b8361e7ff4b89e42ec6cf2859e12e97ec7cd46d80f3e670b51b0751b"} Oct 10 10:06:51 crc kubenswrapper[4669]: I1010 10:06:51.861746 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" podStartSLOduration=3.861724278 podStartE2EDuration="3.861724278s" podCreationTimestamp="2025-10-10 10:06:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:06:51.851418198 +0000 UTC m=+3354.867436950" watchObservedRunningTime="2025-10-10 10:06:51.861724278 +0000 UTC m=+3354.877743020" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.240448 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.256533 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.283781 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-config-data\") pod \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.283887 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-horizon-secret-key\") pod \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.284063 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-scripts\") pod \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.284097 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-logs\") pod \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.284139 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfnl8\" (UniqueName: \"kubernetes.io/projected/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-kube-api-access-dfnl8\") pod \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\" (UID: \"3d1deb67-bb78-4a89-9561-4e916e5e9ad4\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.289282 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-logs" (OuterVolumeSpecName: "logs") pod "3d1deb67-bb78-4a89-9561-4e916e5e9ad4" (UID: "3d1deb67-bb78-4a89-9561-4e916e5e9ad4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.300139 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3d1deb67-bb78-4a89-9561-4e916e5e9ad4" (UID: "3d1deb67-bb78-4a89-9561-4e916e5e9ad4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.304388 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-kube-api-access-dfnl8" (OuterVolumeSpecName: "kube-api-access-dfnl8") pod "3d1deb67-bb78-4a89-9561-4e916e5e9ad4" (UID: "3d1deb67-bb78-4a89-9561-4e916e5e9ad4"). InnerVolumeSpecName "kube-api-access-dfnl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.392045 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-logs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.392074 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfnl8\" (UniqueName: \"kubernetes.io/projected/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-kube-api-access-dfnl8\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.392085 4669 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.424920 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-config-data" (OuterVolumeSpecName: "config-data") pod "3d1deb67-bb78-4a89-9561-4e916e5e9ad4" (UID: "3d1deb67-bb78-4a89-9561-4e916e5e9ad4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.448571 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.460952 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-scripts" (OuterVolumeSpecName: "scripts") pod "3d1deb67-bb78-4a89-9561-4e916e5e9ad4" (UID: "3d1deb67-bb78-4a89-9561-4e916e5e9ad4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.494433 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-scripts\") pod \"d8cd5531-5b53-4fe7-a052-e0229353184d\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.494549 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8cd5531-5b53-4fe7-a052-e0229353184d-horizon-secret-key\") pod \"d8cd5531-5b53-4fe7-a052-e0229353184d\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.494650 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8cd5531-5b53-4fe7-a052-e0229353184d-logs\") pod \"d8cd5531-5b53-4fe7-a052-e0229353184d\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.494729 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrqrx\" (UniqueName: \"kubernetes.io/projected/d8cd5531-5b53-4fe7-a052-e0229353184d-kube-api-access-qrqrx\") pod \"d8cd5531-5b53-4fe7-a052-e0229353184d\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.494747 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-config-data\") pod \"d8cd5531-5b53-4fe7-a052-e0229353184d\" (UID: \"d8cd5531-5b53-4fe7-a052-e0229353184d\") " Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.495249 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.495262 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d1deb67-bb78-4a89-9561-4e916e5e9ad4-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.497040 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8cd5531-5b53-4fe7-a052-e0229353184d-logs" (OuterVolumeSpecName: "logs") pod "d8cd5531-5b53-4fe7-a052-e0229353184d" (UID: "d8cd5531-5b53-4fe7-a052-e0229353184d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.539833 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8cd5531-5b53-4fe7-a052-e0229353184d-kube-api-access-qrqrx" (OuterVolumeSpecName: "kube-api-access-qrqrx") pod "d8cd5531-5b53-4fe7-a052-e0229353184d" (UID: "d8cd5531-5b53-4fe7-a052-e0229353184d"). InnerVolumeSpecName "kube-api-access-qrqrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.569305 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-config-data" (OuterVolumeSpecName: "config-data") pod "d8cd5531-5b53-4fe7-a052-e0229353184d" (UID: "d8cd5531-5b53-4fe7-a052-e0229353184d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.575383 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cd5531-5b53-4fe7-a052-e0229353184d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d8cd5531-5b53-4fe7-a052-e0229353184d" (UID: "d8cd5531-5b53-4fe7-a052-e0229353184d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.590175 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-scripts" (OuterVolumeSpecName: "scripts") pod "d8cd5531-5b53-4fe7-a052-e0229353184d" (UID: "d8cd5531-5b53-4fe7-a052-e0229353184d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.604656 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8cd5531-5b53-4fe7-a052-e0229353184d-logs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.604684 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrqrx\" (UniqueName: \"kubernetes.io/projected/d8cd5531-5b53-4fe7-a052-e0229353184d-kube-api-access-qrqrx\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.604695 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.604704 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8cd5531-5b53-4fe7-a052-e0229353184d-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.604712 4669 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8cd5531-5b53-4fe7-a052-e0229353184d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.878621 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bcdd76bc7-s8ksq" event={"ID":"3d1deb67-bb78-4a89-9561-4e916e5e9ad4","Type":"ContainerDied","Data":"a6e76c17a369e91e49cd820495e125ccc8a31acae5e1ef71e20817c4e0f01abf"} Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.878673 4669 scope.go:117] "RemoveContainer" containerID="8bd1daf033e5f0a1152d540c79ded0c7e56f4f58de2643d12a7501818072dbfe" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.878805 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bcdd76bc7-s8ksq" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.887425 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"dfb11810-489a-4b3b-9ae2-63fc4081c8fa","Type":"ContainerStarted","Data":"494c9066e91379b9a01fb082fe2a124114c9d3e0b4a96e02dbad4bde837bb33a"} Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.887733 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.903571 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e5258eea-9c8f-4551-9874-af8c46837ec2","Type":"ContainerStarted","Data":"82cfe38b8df5ed4f0e7ea275b31671a603229b0138c350c7fb4a589593d0fa43"} Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.903635 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e5258eea-9c8f-4551-9874-af8c46837ec2","Type":"ContainerStarted","Data":"018731122a86fe31ad05fa72b87b05005bb95f79f2f0bdefdb7a018a035c2027"} Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.919512 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56bb6b7c6f-x4hxx" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.919518 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.919502764 podStartE2EDuration="4.919502764s" podCreationTimestamp="2025-10-10 10:06:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:06:52.910180255 +0000 UTC m=+3355.926198997" watchObservedRunningTime="2025-10-10 10:06:52.919502764 +0000 UTC m=+3355.935521506" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.919667 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56bb6b7c6f-x4hxx" event={"ID":"d8cd5531-5b53-4fe7-a052-e0229353184d","Type":"ContainerDied","Data":"faec2d9f0ce6b72e3fcd1ee6866963bd6b371bf0c5f129f28638f39bcbf7e652"} Oct 10 10:06:52 crc kubenswrapper[4669]: E1010 10:06:52.943870 4669 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d1deb67_bb78_4a89_9561_4e916e5e9ad4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d1deb67_bb78_4a89_9561_4e916e5e9ad4.slice/crio-a6e76c17a369e91e49cd820495e125ccc8a31acae5e1ef71e20817c4e0f01abf\": RecentStats: unable to find data in memory cache]" Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.950693 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bcdd76bc7-s8ksq"] Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.968375 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bcdd76bc7-s8ksq"] Oct 10 10:06:52 crc kubenswrapper[4669]: I1010 10:06:52.986323 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=5.015498956 podStartE2EDuration="5.986299884s" podCreationTimestamp="2025-10-10 10:06:47 +0000 UTC" firstStartedPulling="2025-10-10 10:06:49.435739974 +0000 UTC m=+3352.451758716" lastFinishedPulling="2025-10-10 10:06:50.406540912 +0000 UTC m=+3353.422559644" observedRunningTime="2025-10-10 10:06:52.969807866 +0000 UTC m=+3355.985826608" watchObservedRunningTime="2025-10-10 10:06:52.986299884 +0000 UTC m=+3356.002318626" Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.002045 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56bb6b7c6f-x4hxx"] Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.010233 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-56bb6b7c6f-x4hxx"] Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.111779 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-566f97fbdd-plcxg" podUID="0040911a-8dff-49e4-99af-aafb84f7639a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.249:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.155510 4669 scope.go:117] "RemoveContainer" containerID="50243f1d5bad56559ea2a44eb8837f9acc033af9e4ff55ea3d9a81249f8fb61a" Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.181790 4669 scope.go:117] "RemoveContainer" containerID="9766feaf40e48542a096684c3ca9b44d3bea50756217264471b768047903fac3" Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.415970 4669 scope.go:117] "RemoveContainer" containerID="a6d58339f9ad409f7dddb3adbbb7f6ffa8664790c3ed4adb5619217c62271fac" Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.806312 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" path="/var/lib/kubelet/pods/3d1deb67-bb78-4a89-9561-4e916e5e9ad4/volumes" Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.807368 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" path="/var/lib/kubelet/pods/d8cd5531-5b53-4fe7-a052-e0229353184d/volumes" Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.936440 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api-log" containerID="cri-o://d05f7c14b8361e7ff4b89e42ec6cf2859e12e97ec7cd46d80f3e670b51b0751b" gracePeriod=30 Oct 10 10:06:53 crc kubenswrapper[4669]: I1010 10:06:53.936964 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api" containerID="cri-o://494c9066e91379b9a01fb082fe2a124114c9d3e0b4a96e02dbad4bde837bb33a" gracePeriod=30 Oct 10 10:06:54 crc kubenswrapper[4669]: I1010 10:06:54.277289 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:06:54 crc kubenswrapper[4669]: I1010 10:06:54.277347 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:06:54 crc kubenswrapper[4669]: I1010 10:06:54.946472 4669 generic.go:334] "Generic (PLEG): container finished" podID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerID="494c9066e91379b9a01fb082fe2a124114c9d3e0b4a96e02dbad4bde837bb33a" exitCode=0 Oct 10 10:06:54 crc kubenswrapper[4669]: I1010 10:06:54.946732 4669 generic.go:334] "Generic (PLEG): container finished" podID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerID="d05f7c14b8361e7ff4b89e42ec6cf2859e12e97ec7cd46d80f3e670b51b0751b" exitCode=143 Oct 10 10:06:54 crc kubenswrapper[4669]: I1010 10:06:54.946548 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"dfb11810-489a-4b3b-9ae2-63fc4081c8fa","Type":"ContainerDied","Data":"494c9066e91379b9a01fb082fe2a124114c9d3e0b4a96e02dbad4bde837bb33a"} Oct 10 10:06:54 crc kubenswrapper[4669]: I1010 10:06:54.946770 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"dfb11810-489a-4b3b-9ae2-63fc4081c8fa","Type":"ContainerDied","Data":"d05f7c14b8361e7ff4b89e42ec6cf2859e12e97ec7cd46d80f3e670b51b0751b"} Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.084788 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.264939 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m29v\" (UniqueName: \"kubernetes.io/projected/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-kube-api-access-4m29v\") pod \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.265021 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-logs\") pod \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.265094 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-combined-ca-bundle\") pod \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.265690 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-logs" (OuterVolumeSpecName: "logs") pod "dfb11810-489a-4b3b-9ae2-63fc4081c8fa" (UID: "dfb11810-489a-4b3b-9ae2-63fc4081c8fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.265773 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-scripts\") pod \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.266649 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data-custom\") pod \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.266710 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-etc-machine-id\") pod \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.266730 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data\") pod \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\" (UID: \"dfb11810-489a-4b3b-9ae2-63fc4081c8fa\") " Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.267494 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-logs\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.270086 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dfb11810-489a-4b3b-9ae2-63fc4081c8fa" (UID: "dfb11810-489a-4b3b-9ae2-63fc4081c8fa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.275916 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-kube-api-access-4m29v" (OuterVolumeSpecName: "kube-api-access-4m29v") pod "dfb11810-489a-4b3b-9ae2-63fc4081c8fa" (UID: "dfb11810-489a-4b3b-9ae2-63fc4081c8fa"). InnerVolumeSpecName "kube-api-access-4m29v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.304941 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-scripts" (OuterVolumeSpecName: "scripts") pod "dfb11810-489a-4b3b-9ae2-63fc4081c8fa" (UID: "dfb11810-489a-4b3b-9ae2-63fc4081c8fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.305077 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dfb11810-489a-4b3b-9ae2-63fc4081c8fa" (UID: "dfb11810-489a-4b3b-9ae2-63fc4081c8fa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.334844 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfb11810-489a-4b3b-9ae2-63fc4081c8fa" (UID: "dfb11810-489a-4b3b-9ae2-63fc4081c8fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.369385 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.369420 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.369429 4669 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.369439 4669 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.369448 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m29v\" (UniqueName: \"kubernetes.io/projected/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-kube-api-access-4m29v\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.393741 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data" (OuterVolumeSpecName: "config-data") pod "dfb11810-489a-4b3b-9ae2-63fc4081c8fa" (UID: "dfb11810-489a-4b3b-9ae2-63fc4081c8fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.471729 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfb11810-489a-4b3b-9ae2-63fc4081c8fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.967687 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"dfb11810-489a-4b3b-9ae2-63fc4081c8fa","Type":"ContainerDied","Data":"0f8ec6cb84fc0798faaea988751350ab1af48198e622ca958fb104f3e885a640"} Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.967720 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.967795 4669 scope.go:117] "RemoveContainer" containerID="494c9066e91379b9a01fb082fe2a124114c9d3e0b4a96e02dbad4bde837bb33a" Oct 10 10:06:55 crc kubenswrapper[4669]: I1010 10:06:55.992101 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.006089 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.018851 4669 scope.go:117] "RemoveContainer" containerID="d05f7c14b8361e7ff4b89e42ec6cf2859e12e97ec7cd46d80f3e670b51b0751b" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.031387 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:56 crc kubenswrapper[4669]: E1010 10:06:56.031874 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.031897 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon" Oct 10 10:06:56 crc kubenswrapper[4669]: E1010 10:06:56.031918 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.031925 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon" Oct 10 10:06:56 crc kubenswrapper[4669]: E1010 10:06:56.031935 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon-log" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.031941 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon-log" Oct 10 10:06:56 crc kubenswrapper[4669]: E1010 10:06:56.031955 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api-log" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.031961 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api-log" Oct 10 10:06:56 crc kubenswrapper[4669]: E1010 10:06:56.031976 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon-log" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.031984 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon-log" Oct 10 10:06:56 crc kubenswrapper[4669]: E1010 10:06:56.031999 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.032006 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.032176 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.032193 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api-log" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.032203 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.032212 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8cd5531-5b53-4fe7-a052-e0229353184d" containerName="horizon-log" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.032224 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d1deb67-bb78-4a89-9561-4e916e5e9ad4" containerName="horizon-log" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.032233 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" containerName="manila-api" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.033356 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.040489 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.046514 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.049891 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.050166 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090041 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-public-tls-certs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090097 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-etc-machine-id\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090120 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090136 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-config-data-custom\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090210 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-logs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090442 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-scripts\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090468 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dcn6\" (UniqueName: \"kubernetes.io/projected/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-kube-api-access-8dcn6\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090502 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-config-data\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.090515 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-internal-tls-certs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.192119 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-config-data\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.192186 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-internal-tls-certs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.192873 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-public-tls-certs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.193156 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-etc-machine-id\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.193188 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.193204 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-config-data-custom\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.193220 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-etc-machine-id\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.195254 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-logs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.195351 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-scripts\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.195387 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dcn6\" (UniqueName: \"kubernetes.io/projected/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-kube-api-access-8dcn6\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.195879 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-logs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.198842 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-internal-tls-certs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.199117 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-config-data\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.206273 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-public-tls-certs\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.220301 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.220788 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-scripts\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.221922 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-config-data-custom\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.232217 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dcn6\" (UniqueName: \"kubernetes.io/projected/53ead5e1-28e3-4677-99e0-e2faa30ae9a5-kube-api-access-8dcn6\") pod \"manila-api-0\" (UID: \"53ead5e1-28e3-4677-99e0-e2faa30ae9a5\") " pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.251618 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.251952 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-central-agent" containerID="cri-o://b40d7bcd632155f9ffca95723093002b995fb03d2736c6b048a1f27e594e39be" gracePeriod=30 Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.252478 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="proxy-httpd" containerID="cri-o://54bfbc40fe297f1b48b18ac6cdffd691ed80efa7f722956a5ba5106dac538192" gracePeriod=30 Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.252537 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="sg-core" containerID="cri-o://710ea080bd15ec03e3840b34e1a2004e345429d935400bdc67e3f5c4e9ff941c" gracePeriod=30 Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.252572 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-notification-agent" containerID="cri-o://7f9bd1b093c85231fbebb5d1df480ec463cd334b04a08c2f0550744d2a6f14e8" gracePeriod=30 Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.352238 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 10:06:56 crc kubenswrapper[4669]: I1010 10:06:56.992347 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 10 10:06:57 crc kubenswrapper[4669]: I1010 10:06:57.010459 4669 generic.go:334] "Generic (PLEG): container finished" podID="386e6beb-5cc7-46d0-a201-008319df489e" containerID="54bfbc40fe297f1b48b18ac6cdffd691ed80efa7f722956a5ba5106dac538192" exitCode=0 Oct 10 10:06:57 crc kubenswrapper[4669]: I1010 10:06:57.010488 4669 generic.go:334] "Generic (PLEG): container finished" podID="386e6beb-5cc7-46d0-a201-008319df489e" containerID="710ea080bd15ec03e3840b34e1a2004e345429d935400bdc67e3f5c4e9ff941c" exitCode=2 Oct 10 10:06:57 crc kubenswrapper[4669]: I1010 10:06:57.010497 4669 generic.go:334] "Generic (PLEG): container finished" podID="386e6beb-5cc7-46d0-a201-008319df489e" containerID="b40d7bcd632155f9ffca95723093002b995fb03d2736c6b048a1f27e594e39be" exitCode=0 Oct 10 10:06:57 crc kubenswrapper[4669]: I1010 10:06:57.010748 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerDied","Data":"54bfbc40fe297f1b48b18ac6cdffd691ed80efa7f722956a5ba5106dac538192"} Oct 10 10:06:57 crc kubenswrapper[4669]: I1010 10:06:57.010853 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerDied","Data":"710ea080bd15ec03e3840b34e1a2004e345429d935400bdc67e3f5c4e9ff941c"} Oct 10 10:06:57 crc kubenswrapper[4669]: I1010 10:06:57.010870 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerDied","Data":"b40d7bcd632155f9ffca95723093002b995fb03d2736c6b048a1f27e594e39be"} Oct 10 10:06:57 crc kubenswrapper[4669]: I1010 10:06:57.846293 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfb11810-489a-4b3b-9ae2-63fc4081c8fa" path="/var/lib/kubelet/pods/dfb11810-489a-4b3b-9ae2-63fc4081c8fa/volumes" Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.022134 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"53ead5e1-28e3-4677-99e0-e2faa30ae9a5","Type":"ContainerStarted","Data":"0ca80babf155ad3e275aa1297dc7fc25a4a64a793ba08cbcc10d79da8b5d1f99"} Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.022470 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"53ead5e1-28e3-4677-99e0-e2faa30ae9a5","Type":"ContainerStarted","Data":"0dd0586464757daca1583b77472a166d898f5b536c89cb0cfe17229f01c662b6"} Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.031250 4669 generic.go:334] "Generic (PLEG): container finished" podID="386e6beb-5cc7-46d0-a201-008319df489e" containerID="7f9bd1b093c85231fbebb5d1df480ec463cd334b04a08c2f0550744d2a6f14e8" exitCode=0 Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.031303 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerDied","Data":"7f9bd1b093c85231fbebb5d1df480ec463cd334b04a08c2f0550744d2a6f14e8"} Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.583503 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.723779 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58d5f4b747-fl954" Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.798721 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d4c9c94c9-dbk8f"] Oct 10 10:06:58 crc kubenswrapper[4669]: I1010 10:06:58.798926 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerName="dnsmasq-dns" containerID="cri-o://7bb695aaeecf2c868512e35ab57bc26f1de3403e4962da54add9988f5584e6be" gracePeriod=10 Oct 10 10:06:59 crc kubenswrapper[4669]: I1010 10:06:59.049752 4669 generic.go:334] "Generic (PLEG): container finished" podID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerID="7bb695aaeecf2c868512e35ab57bc26f1de3403e4962da54add9988f5584e6be" exitCode=0 Oct 10 10:06:59 crc kubenswrapper[4669]: I1010 10:06:59.049793 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" event={"ID":"7bc42ad9-e7d1-4579-ab07-f0b625372429","Type":"ContainerDied","Data":"7bb695aaeecf2c868512e35ab57bc26f1de3403e4962da54add9988f5584e6be"} Oct 10 10:06:59 crc kubenswrapper[4669]: I1010 10:06:59.790283 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:07:00 crc kubenswrapper[4669]: I1010 10:07:00.020193 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:07:01 crc kubenswrapper[4669]: I1010 10:07:01.034562 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.196:5353: connect: connection refused" Oct 10 10:07:01 crc kubenswrapper[4669]: I1010 10:07:01.602239 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:07:01 crc kubenswrapper[4669]: I1010 10:07:01.954890 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-566f97fbdd-plcxg" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.097298 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7944cbccd6-fnqbk"] Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.128601 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7944cbccd6-fnqbk" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon-log" containerID="cri-o://5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65" gracePeriod=30 Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.128890 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7944cbccd6-fnqbk" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" containerID="cri-o://c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298" gracePeriod=30 Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.286732 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.329242 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393234 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-dns-svc\") pod \"7bc42ad9-e7d1-4579-ab07-f0b625372429\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393285 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqzkt\" (UniqueName: \"kubernetes.io/projected/7bc42ad9-e7d1-4579-ab07-f0b625372429-kube-api-access-wqzkt\") pod \"7bc42ad9-e7d1-4579-ab07-f0b625372429\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393423 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-run-httpd\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393466 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-config-data\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393493 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-config\") pod \"7bc42ad9-e7d1-4579-ab07-f0b625372429\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393516 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-nb\") pod \"7bc42ad9-e7d1-4579-ab07-f0b625372429\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393532 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-combined-ca-bundle\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393613 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-log-httpd\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393639 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-scripts\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393682 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-ceilometer-tls-certs\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393743 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-sg-core-conf-yaml\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393766 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8txcp\" (UniqueName: \"kubernetes.io/projected/386e6beb-5cc7-46d0-a201-008319df489e-kube-api-access-8txcp\") pod \"386e6beb-5cc7-46d0-a201-008319df489e\" (UID: \"386e6beb-5cc7-46d0-a201-008319df489e\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393784 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-openstack-edpm-ipam\") pod \"7bc42ad9-e7d1-4579-ab07-f0b625372429\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.393805 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-sb\") pod \"7bc42ad9-e7d1-4579-ab07-f0b625372429\" (UID: \"7bc42ad9-e7d1-4579-ab07-f0b625372429\") " Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.395654 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.418041 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.418341 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/386e6beb-5cc7-46d0-a201-008319df489e-kube-api-access-8txcp" (OuterVolumeSpecName: "kube-api-access-8txcp") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "kube-api-access-8txcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.428605 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-scripts" (OuterVolumeSpecName: "scripts") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.431076 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc42ad9-e7d1-4579-ab07-f0b625372429-kube-api-access-wqzkt" (OuterVolumeSpecName: "kube-api-access-wqzkt") pod "7bc42ad9-e7d1-4579-ab07-f0b625372429" (UID: "7bc42ad9-e7d1-4579-ab07-f0b625372429"). InnerVolumeSpecName "kube-api-access-wqzkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.486812 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.496695 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.496729 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8txcp\" (UniqueName: \"kubernetes.io/projected/386e6beb-5cc7-46d0-a201-008319df489e-kube-api-access-8txcp\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.496738 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqzkt\" (UniqueName: \"kubernetes.io/projected/7bc42ad9-e7d1-4579-ab07-f0b625372429-kube-api-access-wqzkt\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.496747 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.496756 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/386e6beb-5cc7-46d0-a201-008319df489e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.496764 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.521885 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7bc42ad9-e7d1-4579-ab07-f0b625372429" (UID: "7bc42ad9-e7d1-4579-ab07-f0b625372429"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.541741 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7bc42ad9-e7d1-4579-ab07-f0b625372429" (UID: "7bc42ad9-e7d1-4579-ab07-f0b625372429"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.561475 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-config" (OuterVolumeSpecName: "config") pod "7bc42ad9-e7d1-4579-ab07-f0b625372429" (UID: "7bc42ad9-e7d1-4579-ab07-f0b625372429"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.584527 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.587288 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7bc42ad9-e7d1-4579-ab07-f0b625372429" (UID: "7bc42ad9-e7d1-4579-ab07-f0b625372429"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.598767 4669 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.598806 4669 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-config\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.598821 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.598832 4669 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.598843 4669 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.616889 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7bc42ad9-e7d1-4579-ab07-f0b625372429" (UID: "7bc42ad9-e7d1-4579-ab07-f0b625372429"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.659488 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.702472 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.702789 4669 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7bc42ad9-e7d1-4579-ab07-f0b625372429-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.722548 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-config-data" (OuterVolumeSpecName: "config-data") pod "386e6beb-5cc7-46d0-a201-008319df489e" (UID: "386e6beb-5cc7-46d0-a201-008319df489e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:02 crc kubenswrapper[4669]: I1010 10:07:02.813047 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386e6beb-5cc7-46d0-a201-008319df489e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.153189 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"53ead5e1-28e3-4677-99e0-e2faa30ae9a5","Type":"ContainerStarted","Data":"e9fb14d268521c2ba549ca2ada605e4adf4c9412d3945953d76d520217d72f85"} Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.154457 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.162306 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.162340 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"386e6beb-5cc7-46d0-a201-008319df489e","Type":"ContainerDied","Data":"3a9553be6248c003a7a80166b2ff2d2ae4cd52ac5e30e5905c507a7f3dc041fa"} Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.162398 4669 scope.go:117] "RemoveContainer" containerID="54bfbc40fe297f1b48b18ac6cdffd691ed80efa7f722956a5ba5106dac538192" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.164768 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" event={"ID":"7bc42ad9-e7d1-4579-ab07-f0b625372429","Type":"ContainerDied","Data":"fea1e8ee6f8d160215f8ee79a78ee52133558a8a9012e926cc5f4617bb021ac1"} Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.164834 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d4c9c94c9-dbk8f" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.167997 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"926f04a8-ae97-4fe4-aaaa-edbec3948f79","Type":"ContainerStarted","Data":"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e"} Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.191010 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=8.190992912 podStartE2EDuration="8.190992912s" podCreationTimestamp="2025-10-10 10:06:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:07:03.188657507 +0000 UTC m=+3366.204676249" watchObservedRunningTime="2025-10-10 10:07:03.190992912 +0000 UTC m=+3366.207011654" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.224283 4669 scope.go:117] "RemoveContainer" containerID="710ea080bd15ec03e3840b34e1a2004e345429d935400bdc67e3f5c4e9ff941c" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.241199 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d4c9c94c9-dbk8f"] Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.280777 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d4c9c94c9-dbk8f"] Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.300866 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.309655 4669 scope.go:117] "RemoveContainer" containerID="7f9bd1b093c85231fbebb5d1df480ec463cd334b04a08c2f0550744d2a6f14e8" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.313136 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.326761 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:03 crc kubenswrapper[4669]: E1010 10:07:03.327269 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="sg-core" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327295 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="sg-core" Oct 10 10:07:03 crc kubenswrapper[4669]: E1010 10:07:03.327311 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="proxy-httpd" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327319 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="proxy-httpd" Oct 10 10:07:03 crc kubenswrapper[4669]: E1010 10:07:03.327332 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-notification-agent" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327341 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-notification-agent" Oct 10 10:07:03 crc kubenswrapper[4669]: E1010 10:07:03.327355 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-central-agent" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327363 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-central-agent" Oct 10 10:07:03 crc kubenswrapper[4669]: E1010 10:07:03.327375 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerName="init" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327380 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerName="init" Oct 10 10:07:03 crc kubenswrapper[4669]: E1010 10:07:03.327398 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerName="dnsmasq-dns" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327404 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerName="dnsmasq-dns" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327641 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="sg-core" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327661 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" containerName="dnsmasq-dns" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327676 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-central-agent" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327687 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="proxy-httpd" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.327708 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="386e6beb-5cc7-46d0-a201-008319df489e" containerName="ceilometer-notification-agent" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.329839 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.332367 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.332605 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.332791 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.337924 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:03 crc kubenswrapper[4669]: E1010 10:07:03.345975 4669 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod386e6beb_5cc7_46d0_a201_008319df489e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod386e6beb_5cc7_46d0_a201_008319df489e.slice/crio-3a9553be6248c003a7a80166b2ff2d2ae4cd52ac5e30e5905c507a7f3dc041fa\": RecentStats: unable to find data in memory cache]" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.366205 4669 scope.go:117] "RemoveContainer" containerID="b40d7bcd632155f9ffca95723093002b995fb03d2736c6b048a1f27e594e39be" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.395776 4669 scope.go:117] "RemoveContainer" containerID="7bb695aaeecf2c868512e35ab57bc26f1de3403e4962da54add9988f5584e6be" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.426192 4669 scope.go:117] "RemoveContainer" containerID="5bec3eae24e2177201f6883216457a655f42fea3f002bf2c850ef14e6fbe3541" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462191 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462279 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462318 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmj4s\" (UniqueName: \"kubernetes.io/projected/f7ede6ae-9943-41e2-89d4-f3851dd0d842-kube-api-access-fmj4s\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462378 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462404 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462441 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-scripts\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462483 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.462565 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-config-data\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.564333 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-config-data\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.564391 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.564436 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.564466 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmj4s\" (UniqueName: \"kubernetes.io/projected/f7ede6ae-9943-41e2-89d4-f3851dd0d842-kube-api-access-fmj4s\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.564509 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.564527 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.564575 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-scripts\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.565335 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.565455 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.565636 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.571072 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.571863 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.572297 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-scripts\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.572461 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.574296 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-config-data\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.587017 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmj4s\" (UniqueName: \"kubernetes.io/projected/f7ede6ae-9943-41e2-89d4-f3851dd0d842-kube-api-access-fmj4s\") pod \"ceilometer-0\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.662294 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.835638 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="386e6beb-5cc7-46d0-a201-008319df489e" path="/var/lib/kubelet/pods/386e6beb-5cc7-46d0-a201-008319df489e/volumes" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.837461 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc42ad9-e7d1-4579-ab07-f0b625372429" path="/var/lib/kubelet/pods/7bc42ad9-e7d1-4579-ab07-f0b625372429/volumes" Oct 10 10:07:03 crc kubenswrapper[4669]: I1010 10:07:03.888374 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:04 crc kubenswrapper[4669]: I1010 10:07:04.183072 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"926f04a8-ae97-4fe4-aaaa-edbec3948f79","Type":"ContainerStarted","Data":"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e"} Oct 10 10:07:04 crc kubenswrapper[4669]: I1010 10:07:04.217278 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.759288056 podStartE2EDuration="17.217243737s" podCreationTimestamp="2025-10-10 10:06:47 +0000 UTC" firstStartedPulling="2025-10-10 10:06:49.471427737 +0000 UTC m=+3352.487446479" lastFinishedPulling="2025-10-10 10:07:01.929383418 +0000 UTC m=+3364.945402160" observedRunningTime="2025-10-10 10:07:04.20735902 +0000 UTC m=+3367.223377762" watchObservedRunningTime="2025-10-10 10:07:04.217243737 +0000 UTC m=+3367.233262479" Oct 10 10:07:04 crc kubenswrapper[4669]: I1010 10:07:04.232926 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:05 crc kubenswrapper[4669]: I1010 10:07:05.191335 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerStarted","Data":"0dfff9d0961d833fd5321d27c7159c0e3d50f4cd7e1aca47cc446d78317873a5"} Oct 10 10:07:06 crc kubenswrapper[4669]: I1010 10:07:06.204224 4669 generic.go:334] "Generic (PLEG): container finished" podID="af0a1408-8b73-4070-b491-74d331bcf567" containerID="c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298" exitCode=0 Oct 10 10:07:06 crc kubenswrapper[4669]: I1010 10:07:06.204722 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7944cbccd6-fnqbk" event={"ID":"af0a1408-8b73-4070-b491-74d331bcf567","Type":"ContainerDied","Data":"c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298"} Oct 10 10:07:06 crc kubenswrapper[4669]: I1010 10:07:06.207507 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerStarted","Data":"cfe0c805c01223ad752701d12a5bbe96de5c40d1d92ac6125da8422cb7636e53"} Oct 10 10:07:07 crc kubenswrapper[4669]: I1010 10:07:07.224217 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerStarted","Data":"3b353bc7c6be1d4fc253ffa39098bdb38680a6a23eddd771648136cf4a61bb66"} Oct 10 10:07:07 crc kubenswrapper[4669]: I1010 10:07:07.224531 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerStarted","Data":"dfcc9b53bea512a3f30c1206648ec32f338cd5db9d67fe6279f79dfaf1c88fdd"} Oct 10 10:07:07 crc kubenswrapper[4669]: I1010 10:07:07.482148 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7944cbccd6-fnqbk" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 10 10:07:08 crc kubenswrapper[4669]: I1010 10:07:08.349509 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 10 10:07:09 crc kubenswrapper[4669]: I1010 10:07:09.253085 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerStarted","Data":"9c2014172c974c4e836671fb0ddb6f089a96e4f1f24f587792fde533a05a045f"} Oct 10 10:07:09 crc kubenswrapper[4669]: I1010 10:07:09.253235 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-central-agent" containerID="cri-o://cfe0c805c01223ad752701d12a5bbe96de5c40d1d92ac6125da8422cb7636e53" gracePeriod=30 Oct 10 10:07:09 crc kubenswrapper[4669]: I1010 10:07:09.253286 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="sg-core" containerID="cri-o://3b353bc7c6be1d4fc253ffa39098bdb38680a6a23eddd771648136cf4a61bb66" gracePeriod=30 Oct 10 10:07:09 crc kubenswrapper[4669]: I1010 10:07:09.253282 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="proxy-httpd" containerID="cri-o://9c2014172c974c4e836671fb0ddb6f089a96e4f1f24f587792fde533a05a045f" gracePeriod=30 Oct 10 10:07:09 crc kubenswrapper[4669]: I1010 10:07:09.253537 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 10:07:09 crc kubenswrapper[4669]: I1010 10:07:09.253302 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-notification-agent" containerID="cri-o://dfcc9b53bea512a3f30c1206648ec32f338cd5db9d67fe6279f79dfaf1c88fdd" gracePeriod=30 Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266082 4669 generic.go:334] "Generic (PLEG): container finished" podID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerID="9c2014172c974c4e836671fb0ddb6f089a96e4f1f24f587792fde533a05a045f" exitCode=0 Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266612 4669 generic.go:334] "Generic (PLEG): container finished" podID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerID="3b353bc7c6be1d4fc253ffa39098bdb38680a6a23eddd771648136cf4a61bb66" exitCode=2 Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266626 4669 generic.go:334] "Generic (PLEG): container finished" podID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerID="dfcc9b53bea512a3f30c1206648ec32f338cd5db9d67fe6279f79dfaf1c88fdd" exitCode=0 Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266639 4669 generic.go:334] "Generic (PLEG): container finished" podID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerID="cfe0c805c01223ad752701d12a5bbe96de5c40d1d92ac6125da8422cb7636e53" exitCode=0 Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266662 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerDied","Data":"9c2014172c974c4e836671fb0ddb6f089a96e4f1f24f587792fde533a05a045f"} Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266692 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerDied","Data":"3b353bc7c6be1d4fc253ffa39098bdb38680a6a23eddd771648136cf4a61bb66"} Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266703 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerDied","Data":"dfcc9b53bea512a3f30c1206648ec32f338cd5db9d67fe6279f79dfaf1c88fdd"} Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266714 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerDied","Data":"cfe0c805c01223ad752701d12a5bbe96de5c40d1d92ac6125da8422cb7636e53"} Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266726 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ede6ae-9943-41e2-89d4-f3851dd0d842","Type":"ContainerDied","Data":"0dfff9d0961d833fd5321d27c7159c0e3d50f4cd7e1aca47cc446d78317873a5"} Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.266739 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dfff9d0961d833fd5321d27c7159c0e3d50f4cd7e1aca47cc446d78317873a5" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.309909 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.340630 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.399228 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418142 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-log-httpd\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418236 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-combined-ca-bundle\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418276 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-config-data\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418306 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-scripts\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418331 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-run-httpd\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418349 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmj4s\" (UniqueName: \"kubernetes.io/projected/f7ede6ae-9943-41e2-89d4-f3851dd0d842-kube-api-access-fmj4s\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418367 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-ceilometer-tls-certs\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418547 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-sg-core-conf-yaml\") pod \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\" (UID: \"f7ede6ae-9943-41e2-89d4-f3851dd0d842\") " Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.418807 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.419214 4669 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.422317 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.425271 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7ede6ae-9943-41e2-89d4-f3851dd0d842-kube-api-access-fmj4s" (OuterVolumeSpecName: "kube-api-access-fmj4s") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "kube-api-access-fmj4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.427434 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-scripts" (OuterVolumeSpecName: "scripts") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.480561 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.481961 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.522289 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.522325 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmj4s\" (UniqueName: \"kubernetes.io/projected/f7ede6ae-9943-41e2-89d4-f3851dd0d842-kube-api-access-fmj4s\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.522334 4669 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ede6ae-9943-41e2-89d4-f3851dd0d842-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.522343 4669 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.522352 4669 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.527904 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.553841 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-config-data" (OuterVolumeSpecName: "config-data") pod "f7ede6ae-9943-41e2-89d4-f3851dd0d842" (UID: "f7ede6ae-9943-41e2-89d4-f3851dd0d842"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.624036 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:10 crc kubenswrapper[4669]: I1010 10:07:10.624082 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ede6ae-9943-41e2-89d4-f3851dd0d842-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.278759 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.279257 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="probe" containerID="cri-o://018731122a86fe31ad05fa72b87b05005bb95f79f2f0bdefdb7a018a035c2027" gracePeriod=30 Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.278964 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="manila-scheduler" containerID="cri-o://82cfe38b8df5ed4f0e7ea275b31671a603229b0138c350c7fb4a589593d0fa43" gracePeriod=30 Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.340949 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.366489 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.392282 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:11 crc kubenswrapper[4669]: E1010 10:07:11.393153 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-notification-agent" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393179 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-notification-agent" Oct 10 10:07:11 crc kubenswrapper[4669]: E1010 10:07:11.393200 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="proxy-httpd" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393207 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="proxy-httpd" Oct 10 10:07:11 crc kubenswrapper[4669]: E1010 10:07:11.393235 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-central-agent" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393241 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-central-agent" Oct 10 10:07:11 crc kubenswrapper[4669]: E1010 10:07:11.393253 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="sg-core" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393262 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="sg-core" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393851 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="sg-core" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393890 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="proxy-httpd" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393906 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-central-agent" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.393928 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" containerName="ceilometer-notification-agent" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.397285 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.401052 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.401193 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.401436 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.425769 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.441324 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.441655 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-scripts\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.441736 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/586ee21d-4f34-48fe-8528-fe925bb8e35a-run-httpd\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.441807 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.441915 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/586ee21d-4f34-48fe-8528-fe925bb8e35a-log-httpd\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.441991 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.442074 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfqm5\" (UniqueName: \"kubernetes.io/projected/586ee21d-4f34-48fe-8528-fe925bb8e35a-kube-api-access-bfqm5\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.442162 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-config-data\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.543792 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.544150 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-scripts\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.544274 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/586ee21d-4f34-48fe-8528-fe925bb8e35a-run-httpd\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.544366 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.544457 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/586ee21d-4f34-48fe-8528-fe925bb8e35a-log-httpd\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.544551 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.544671 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfqm5\" (UniqueName: \"kubernetes.io/projected/586ee21d-4f34-48fe-8528-fe925bb8e35a-kube-api-access-bfqm5\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.544782 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-config-data\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.546864 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/586ee21d-4f34-48fe-8528-fe925bb8e35a-log-httpd\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.546908 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/586ee21d-4f34-48fe-8528-fe925bb8e35a-run-httpd\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.549710 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.550486 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.551411 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-config-data\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.551515 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.556433 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/586ee21d-4f34-48fe-8528-fe925bb8e35a-scripts\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.562429 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfqm5\" (UniqueName: \"kubernetes.io/projected/586ee21d-4f34-48fe-8528-fe925bb8e35a-kube-api-access-bfqm5\") pod \"ceilometer-0\" (UID: \"586ee21d-4f34-48fe-8528-fe925bb8e35a\") " pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.733323 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 10:07:11 crc kubenswrapper[4669]: I1010 10:07:11.807513 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7ede6ae-9943-41e2-89d4-f3851dd0d842" path="/var/lib/kubelet/pods/f7ede6ae-9943-41e2-89d4-f3851dd0d842/volumes" Oct 10 10:07:12 crc kubenswrapper[4669]: I1010 10:07:12.236094 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 10:07:12 crc kubenswrapper[4669]: W1010 10:07:12.251373 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod586ee21d_4f34_48fe_8528_fe925bb8e35a.slice/crio-a42b61d14a004b3d14ea78806922575625389546858a3bb24f97f07cd5d6b46b WatchSource:0}: Error finding container a42b61d14a004b3d14ea78806922575625389546858a3bb24f97f07cd5d6b46b: Status 404 returned error can't find the container with id a42b61d14a004b3d14ea78806922575625389546858a3bb24f97f07cd5d6b46b Oct 10 10:07:12 crc kubenswrapper[4669]: I1010 10:07:12.288916 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"586ee21d-4f34-48fe-8528-fe925bb8e35a","Type":"ContainerStarted","Data":"a42b61d14a004b3d14ea78806922575625389546858a3bb24f97f07cd5d6b46b"} Oct 10 10:07:12 crc kubenswrapper[4669]: I1010 10:07:12.294076 4669 generic.go:334] "Generic (PLEG): container finished" podID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerID="018731122a86fe31ad05fa72b87b05005bb95f79f2f0bdefdb7a018a035c2027" exitCode=0 Oct 10 10:07:12 crc kubenswrapper[4669]: I1010 10:07:12.294110 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e5258eea-9c8f-4551-9874-af8c46837ec2","Type":"ContainerDied","Data":"018731122a86fe31ad05fa72b87b05005bb95f79f2f0bdefdb7a018a035c2027"} Oct 10 10:07:13 crc kubenswrapper[4669]: I1010 10:07:13.304337 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"586ee21d-4f34-48fe-8528-fe925bb8e35a","Type":"ContainerStarted","Data":"d0686d3eb1d1e39faef613015a3f07fe5c86d0dd4b89fd52710190fcf193a65d"} Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.325901 4669 generic.go:334] "Generic (PLEG): container finished" podID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerID="82cfe38b8df5ed4f0e7ea275b31671a603229b0138c350c7fb4a589593d0fa43" exitCode=0 Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.326229 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e5258eea-9c8f-4551-9874-af8c46837ec2","Type":"ContainerDied","Data":"82cfe38b8df5ed4f0e7ea275b31671a603229b0138c350c7fb4a589593d0fa43"} Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.332646 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"586ee21d-4f34-48fe-8528-fe925bb8e35a","Type":"ContainerStarted","Data":"d1c1d9714ff5d9a4a664324ed340c28601c280c2459ba694cc92bcf622cf245c"} Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.540120 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.711263 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dxxt\" (UniqueName: \"kubernetes.io/projected/e5258eea-9c8f-4551-9874-af8c46837ec2-kube-api-access-5dxxt\") pod \"e5258eea-9c8f-4551-9874-af8c46837ec2\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.711992 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-scripts\") pod \"e5258eea-9c8f-4551-9874-af8c46837ec2\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.712130 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-combined-ca-bundle\") pod \"e5258eea-9c8f-4551-9874-af8c46837ec2\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.712265 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data-custom\") pod \"e5258eea-9c8f-4551-9874-af8c46837ec2\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.712466 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data\") pod \"e5258eea-9c8f-4551-9874-af8c46837ec2\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.712613 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5258eea-9c8f-4551-9874-af8c46837ec2-etc-machine-id\") pod \"e5258eea-9c8f-4551-9874-af8c46837ec2\" (UID: \"e5258eea-9c8f-4551-9874-af8c46837ec2\") " Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.713802 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5258eea-9c8f-4551-9874-af8c46837ec2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e5258eea-9c8f-4551-9874-af8c46837ec2" (UID: "e5258eea-9c8f-4551-9874-af8c46837ec2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.721068 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-scripts" (OuterVolumeSpecName: "scripts") pod "e5258eea-9c8f-4551-9874-af8c46837ec2" (UID: "e5258eea-9c8f-4551-9874-af8c46837ec2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.721103 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5258eea-9c8f-4551-9874-af8c46837ec2-kube-api-access-5dxxt" (OuterVolumeSpecName: "kube-api-access-5dxxt") pod "e5258eea-9c8f-4551-9874-af8c46837ec2" (UID: "e5258eea-9c8f-4551-9874-af8c46837ec2"). InnerVolumeSpecName "kube-api-access-5dxxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.722987 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e5258eea-9c8f-4551-9874-af8c46837ec2" (UID: "e5258eea-9c8f-4551-9874-af8c46837ec2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.787156 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5258eea-9c8f-4551-9874-af8c46837ec2" (UID: "e5258eea-9c8f-4551-9874-af8c46837ec2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.817572 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dxxt\" (UniqueName: \"kubernetes.io/projected/e5258eea-9c8f-4551-9874-af8c46837ec2-kube-api-access-5dxxt\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.817800 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.817898 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.817967 4669 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.818037 4669 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e5258eea-9c8f-4551-9874-af8c46837ec2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.849400 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data" (OuterVolumeSpecName: "config-data") pod "e5258eea-9c8f-4551-9874-af8c46837ec2" (UID: "e5258eea-9c8f-4551-9874-af8c46837ec2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:14 crc kubenswrapper[4669]: I1010 10:07:14.919513 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5258eea-9c8f-4551-9874-af8c46837ec2-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.341712 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"586ee21d-4f34-48fe-8528-fe925bb8e35a","Type":"ContainerStarted","Data":"b921496610c8bc4b7cd0363108787e2ed94d66d759f809ec9b074c3f6de51fdc"} Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.345129 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"e5258eea-9c8f-4551-9874-af8c46837ec2","Type":"ContainerDied","Data":"8818281a9fe7920d91feafd13536aa2458f8dc119161ba13315a60eeae155189"} Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.345179 4669 scope.go:117] "RemoveContainer" containerID="018731122a86fe31ad05fa72b87b05005bb95f79f2f0bdefdb7a018a035c2027" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.345198 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.386290 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.386883 4669 scope.go:117] "RemoveContainer" containerID="82cfe38b8df5ed4f0e7ea275b31671a603229b0138c350c7fb4a589593d0fa43" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.417964 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.430420 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:07:15 crc kubenswrapper[4669]: E1010 10:07:15.433697 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="probe" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.433925 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="probe" Oct 10 10:07:15 crc kubenswrapper[4669]: E1010 10:07:15.433957 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="manila-scheduler" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.434207 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="manila-scheduler" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.435134 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="probe" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.435405 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" containerName="manila-scheduler" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.441119 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.441442 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.443967 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.531131 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf826\" (UniqueName: \"kubernetes.io/projected/8a1791d4-98d8-4eac-9676-e409ac975611-kube-api-access-xf826\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.531172 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.531221 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a1791d4-98d8-4eac-9676-e409ac975611-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.531273 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-config-data\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.531344 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-scripts\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.531401 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.633328 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-scripts\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.633442 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.634143 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf826\" (UniqueName: \"kubernetes.io/projected/8a1791d4-98d8-4eac-9676-e409ac975611-kube-api-access-xf826\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.634198 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.634249 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a1791d4-98d8-4eac-9676-e409ac975611-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.634327 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-config-data\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.634403 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a1791d4-98d8-4eac-9676-e409ac975611-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.646651 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.646699 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.647387 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-config-data\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.648680 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a1791d4-98d8-4eac-9676-e409ac975611-scripts\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.665099 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf826\" (UniqueName: \"kubernetes.io/projected/8a1791d4-98d8-4eac-9676-e409ac975611-kube-api-access-xf826\") pod \"manila-scheduler-0\" (UID: \"8a1791d4-98d8-4eac-9676-e409ac975611\") " pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.756393 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 10:07:15 crc kubenswrapper[4669]: I1010 10:07:15.808907 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5258eea-9c8f-4551-9874-af8c46837ec2" path="/var/lib/kubelet/pods/e5258eea-9c8f-4551-9874-af8c46837ec2/volumes" Oct 10 10:07:16 crc kubenswrapper[4669]: I1010 10:07:16.246351 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 10:07:16 crc kubenswrapper[4669]: I1010 10:07:16.362729 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8a1791d4-98d8-4eac-9676-e409ac975611","Type":"ContainerStarted","Data":"e4ae297b7a60e4976050eaeb059bb25a05892b101091e52a665ce0d98fd90c1e"} Oct 10 10:07:16 crc kubenswrapper[4669]: I1010 10:07:16.366790 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"586ee21d-4f34-48fe-8528-fe925bb8e35a","Type":"ContainerStarted","Data":"363a3fa3b964d7eb5f5d14a0b397bb6416cacc1850a3791d807b3cac9769d88f"} Oct 10 10:07:16 crc kubenswrapper[4669]: I1010 10:07:16.367049 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 10:07:16 crc kubenswrapper[4669]: I1010 10:07:16.390236 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.669732576 podStartE2EDuration="5.390218389s" podCreationTimestamp="2025-10-10 10:07:11 +0000 UTC" firstStartedPulling="2025-10-10 10:07:12.271426157 +0000 UTC m=+3375.287444899" lastFinishedPulling="2025-10-10 10:07:15.99191197 +0000 UTC m=+3379.007930712" observedRunningTime="2025-10-10 10:07:16.387820283 +0000 UTC m=+3379.403839025" watchObservedRunningTime="2025-10-10 10:07:16.390218389 +0000 UTC m=+3379.406237131" Oct 10 10:07:17 crc kubenswrapper[4669]: I1010 10:07:17.377827 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8a1791d4-98d8-4eac-9676-e409ac975611","Type":"ContainerStarted","Data":"4e23617d04578b9dfc3cc5c06470890a79e3146bdeb2e90a1aab970a0b492620"} Oct 10 10:07:17 crc kubenswrapper[4669]: I1010 10:07:17.378444 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8a1791d4-98d8-4eac-9676-e409ac975611","Type":"ContainerStarted","Data":"4a3a70689b7b77f2500889d66890fecd64c8acf2d6569b054f529b382fe3dade"} Oct 10 10:07:17 crc kubenswrapper[4669]: I1010 10:07:17.399191 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.39917148 podStartE2EDuration="2.39917148s" podCreationTimestamp="2025-10-10 10:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:07:17.394154879 +0000 UTC m=+3380.410173621" watchObservedRunningTime="2025-10-10 10:07:17.39917148 +0000 UTC m=+3380.415190222" Oct 10 10:07:17 crc kubenswrapper[4669]: I1010 10:07:17.481724 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7944cbccd6-fnqbk" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 10 10:07:17 crc kubenswrapper[4669]: I1010 10:07:17.851060 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 10 10:07:20 crc kubenswrapper[4669]: I1010 10:07:20.027259 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 10 10:07:20 crc kubenswrapper[4669]: I1010 10:07:20.107463 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:07:20 crc kubenswrapper[4669]: I1010 10:07:20.404205 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="probe" containerID="cri-o://da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e" gracePeriod=30 Oct 10 10:07:20 crc kubenswrapper[4669]: I1010 10:07:20.404665 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="manila-share" containerID="cri-o://d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e" gracePeriod=30 Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.356076 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.429508 4669 generic.go:334] "Generic (PLEG): container finished" podID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerID="da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e" exitCode=0 Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.429605 4669 generic.go:334] "Generic (PLEG): container finished" podID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerID="d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e" exitCode=1 Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.429652 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"926f04a8-ae97-4fe4-aaaa-edbec3948f79","Type":"ContainerDied","Data":"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e"} Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.429694 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"926f04a8-ae97-4fe4-aaaa-edbec3948f79","Type":"ContainerDied","Data":"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e"} Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.429714 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"926f04a8-ae97-4fe4-aaaa-edbec3948f79","Type":"ContainerDied","Data":"db5622e287588f7b75837593953a73674b9837a8bf3d1a9d73374f403f6b74c1"} Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.429740 4669 scope.go:117] "RemoveContainer" containerID="da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.429932 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.451987 4669 scope.go:117] "RemoveContainer" containerID="d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.472483 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-var-lib-manila\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.472629 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-scripts\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.472712 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cdls\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-kube-api-access-5cdls\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.472859 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-etc-machine-id\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.472966 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-combined-ca-bundle\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.473124 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data-custom\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.473122 4669 scope.go:117] "RemoveContainer" containerID="da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.473200 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-ceph\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.473235 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data\") pod \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\" (UID: \"926f04a8-ae97-4fe4-aaaa-edbec3948f79\") " Oct 10 10:07:21 crc kubenswrapper[4669]: E1010 10:07:21.477404 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e\": container with ID starting with da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e not found: ID does not exist" containerID="da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.477565 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e"} err="failed to get container status \"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e\": rpc error: code = NotFound desc = could not find container \"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e\": container with ID starting with da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e not found: ID does not exist" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.477608 4669 scope.go:117] "RemoveContainer" containerID="d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.478085 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: E1010 10:07:21.478672 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e\": container with ID starting with d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e not found: ID does not exist" containerID="d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.478723 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e"} err="failed to get container status \"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e\": rpc error: code = NotFound desc = could not find container \"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e\": container with ID starting with d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e not found: ID does not exist" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.478751 4669 scope.go:117] "RemoveContainer" containerID="da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.479040 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-kube-api-access-5cdls" (OuterVolumeSpecName: "kube-api-access-5cdls") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "kube-api-access-5cdls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.479105 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.479196 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e"} err="failed to get container status \"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e\": rpc error: code = NotFound desc = could not find container \"da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e\": container with ID starting with da450103dca70403b7be2d0210f892b3762e871b23dc0e7c444dc03e31806f8e not found: ID does not exist" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.479300 4669 scope.go:117] "RemoveContainer" containerID="d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.479759 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-scripts" (OuterVolumeSpecName: "scripts") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.479917 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e"} err="failed to get container status \"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e\": rpc error: code = NotFound desc = could not find container \"d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e\": container with ID starting with d7c975cb186d9a28744e57e2517dcfcad1966b8afb07f0959effbab24809d77e not found: ID does not exist" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.485632 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.492726 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-ceph" (OuterVolumeSpecName: "ceph") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.536412 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.576289 4669 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.576331 4669 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.576344 4669 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.576356 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.576367 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cdls\" (UniqueName: \"kubernetes.io/projected/926f04a8-ae97-4fe4-aaaa-edbec3948f79-kube-api-access-5cdls\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.576380 4669 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/926f04a8-ae97-4fe4-aaaa-edbec3948f79-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.576391 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.616760 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data" (OuterVolumeSpecName: "config-data") pod "926f04a8-ae97-4fe4-aaaa-edbec3948f79" (UID: "926f04a8-ae97-4fe4-aaaa-edbec3948f79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.677832 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/926f04a8-ae97-4fe4-aaaa-edbec3948f79-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.791808 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.810674 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.817391 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:07:21 crc kubenswrapper[4669]: E1010 10:07:21.817910 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="probe" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.817931 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="probe" Oct 10 10:07:21 crc kubenswrapper[4669]: E1010 10:07:21.817957 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="manila-share" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.817966 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="manila-share" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.818175 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="probe" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.818203 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" containerName="manila-share" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.819317 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.823006 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.848041 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880318 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-ceph\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880368 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-scripts\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880414 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h82t\" (UniqueName: \"kubernetes.io/projected/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-kube-api-access-6h82t\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880430 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-config-data\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880466 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880621 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880801 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.880825 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983219 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h82t\" (UniqueName: \"kubernetes.io/projected/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-kube-api-access-6h82t\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983271 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-config-data\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983303 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983364 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983448 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983463 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983507 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-ceph\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983537 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-scripts\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983649 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.983750 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.987260 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.988721 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-scripts\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.989201 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-config-data\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.991033 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-ceph\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:21 crc kubenswrapper[4669]: I1010 10:07:21.995234 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:22 crc kubenswrapper[4669]: I1010 10:07:22.003184 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h82t\" (UniqueName: \"kubernetes.io/projected/e4865c42-8dbc-4d96-afb9-e65d85f8ec96-kube-api-access-6h82t\") pod \"manila-share-share1-0\" (UID: \"e4865c42-8dbc-4d96-afb9-e65d85f8ec96\") " pod="openstack/manila-share-share1-0" Oct 10 10:07:22 crc kubenswrapper[4669]: I1010 10:07:22.142850 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 10:07:22 crc kubenswrapper[4669]: I1010 10:07:22.707669 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 10:07:23 crc kubenswrapper[4669]: I1010 10:07:23.459175 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e4865c42-8dbc-4d96-afb9-e65d85f8ec96","Type":"ContainerStarted","Data":"86133f4da973f668c40de0c1815171c2d41dbaa89ec33799b7fee0b59b4f7a15"} Oct 10 10:07:23 crc kubenswrapper[4669]: I1010 10:07:23.459433 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e4865c42-8dbc-4d96-afb9-e65d85f8ec96","Type":"ContainerStarted","Data":"a216c1ed89051c60b297d7d30f2808c244fc669565af0fd4b1c858e3eeb38e3b"} Oct 10 10:07:23 crc kubenswrapper[4669]: I1010 10:07:23.808340 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="926f04a8-ae97-4fe4-aaaa-edbec3948f79" path="/var/lib/kubelet/pods/926f04a8-ae97-4fe4-aaaa-edbec3948f79/volumes" Oct 10 10:07:24 crc kubenswrapper[4669]: I1010 10:07:24.274835 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:07:24 crc kubenswrapper[4669]: I1010 10:07:24.274904 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:07:24 crc kubenswrapper[4669]: I1010 10:07:24.471215 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"e4865c42-8dbc-4d96-afb9-e65d85f8ec96","Type":"ContainerStarted","Data":"4e627c7d7342f2f7f64eb971f029c079689debb6069da49426b70f2aece62c1d"} Oct 10 10:07:24 crc kubenswrapper[4669]: I1010 10:07:24.496617 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.496570331 podStartE2EDuration="3.496570331s" podCreationTimestamp="2025-10-10 10:07:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:07:24.489930008 +0000 UTC m=+3387.505948860" watchObservedRunningTime="2025-10-10 10:07:24.496570331 +0000 UTC m=+3387.512589073" Oct 10 10:07:25 crc kubenswrapper[4669]: I1010 10:07:25.756653 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 10 10:07:27 crc kubenswrapper[4669]: I1010 10:07:27.481325 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7944cbccd6-fnqbk" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 10 10:07:27 crc kubenswrapper[4669]: I1010 10:07:27.481780 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.143759 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.548338 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.560795 4669 generic.go:334] "Generic (PLEG): container finished" podID="af0a1408-8b73-4070-b491-74d331bcf567" containerID="5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65" exitCode=137 Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.560846 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7944cbccd6-fnqbk" event={"ID":"af0a1408-8b73-4070-b491-74d331bcf567","Type":"ContainerDied","Data":"5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65"} Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.560876 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7944cbccd6-fnqbk" event={"ID":"af0a1408-8b73-4070-b491-74d331bcf567","Type":"ContainerDied","Data":"60d4dd2fdafab6673200c8a06a60c01c66092828b1193d8c2970ede1aeae985d"} Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.560895 4669 scope.go:117] "RemoveContainer" containerID="c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.561035 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7944cbccd6-fnqbk" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.720421 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-scripts\") pod \"af0a1408-8b73-4070-b491-74d331bcf567\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.720495 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-combined-ca-bundle\") pod \"af0a1408-8b73-4070-b491-74d331bcf567\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.720552 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-config-data\") pod \"af0a1408-8b73-4070-b491-74d331bcf567\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.721047 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-tls-certs\") pod \"af0a1408-8b73-4070-b491-74d331bcf567\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.721155 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0a1408-8b73-4070-b491-74d331bcf567-logs\") pod \"af0a1408-8b73-4070-b491-74d331bcf567\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.721331 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6wwp\" (UniqueName: \"kubernetes.io/projected/af0a1408-8b73-4070-b491-74d331bcf567-kube-api-access-h6wwp\") pod \"af0a1408-8b73-4070-b491-74d331bcf567\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.721396 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-secret-key\") pod \"af0a1408-8b73-4070-b491-74d331bcf567\" (UID: \"af0a1408-8b73-4070-b491-74d331bcf567\") " Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.721858 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af0a1408-8b73-4070-b491-74d331bcf567-logs" (OuterVolumeSpecName: "logs") pod "af0a1408-8b73-4070-b491-74d331bcf567" (UID: "af0a1408-8b73-4070-b491-74d331bcf567"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.722290 4669 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0a1408-8b73-4070-b491-74d331bcf567-logs\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.728393 4669 scope.go:117] "RemoveContainer" containerID="5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.728492 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0a1408-8b73-4070-b491-74d331bcf567-kube-api-access-h6wwp" (OuterVolumeSpecName: "kube-api-access-h6wwp") pod "af0a1408-8b73-4070-b491-74d331bcf567" (UID: "af0a1408-8b73-4070-b491-74d331bcf567"). InnerVolumeSpecName "kube-api-access-h6wwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.741749 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "af0a1408-8b73-4070-b491-74d331bcf567" (UID: "af0a1408-8b73-4070-b491-74d331bcf567"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.750554 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-scripts" (OuterVolumeSpecName: "scripts") pod "af0a1408-8b73-4070-b491-74d331bcf567" (UID: "af0a1408-8b73-4070-b491-74d331bcf567"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.758825 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af0a1408-8b73-4070-b491-74d331bcf567" (UID: "af0a1408-8b73-4070-b491-74d331bcf567"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.767143 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-config-data" (OuterVolumeSpecName: "config-data") pod "af0a1408-8b73-4070-b491-74d331bcf567" (UID: "af0a1408-8b73-4070-b491-74d331bcf567"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.779051 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "af0a1408-8b73-4070-b491-74d331bcf567" (UID: "af0a1408-8b73-4070-b491-74d331bcf567"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.834477 4669 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.834529 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6wwp\" (UniqueName: \"kubernetes.io/projected/af0a1408-8b73-4070-b491-74d331bcf567-kube-api-access-h6wwp\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.834545 4669 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.834558 4669 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.834571 4669 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0a1408-8b73-4070-b491-74d331bcf567-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.834602 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af0a1408-8b73-4070-b491-74d331bcf567-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.836449 4669 scope.go:117] "RemoveContainer" containerID="c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298" Oct 10 10:07:32 crc kubenswrapper[4669]: E1010 10:07:32.836922 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298\": container with ID starting with c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298 not found: ID does not exist" containerID="c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.836974 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298"} err="failed to get container status \"c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298\": rpc error: code = NotFound desc = could not find container \"c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298\": container with ID starting with c15a98d0f9564aa580451d28b4e40925b561755f8830c45d239d8befc3064298 not found: ID does not exist" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.837008 4669 scope.go:117] "RemoveContainer" containerID="5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65" Oct 10 10:07:32 crc kubenswrapper[4669]: E1010 10:07:32.837417 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65\": container with ID starting with 5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65 not found: ID does not exist" containerID="5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.837449 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65"} err="failed to get container status \"5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65\": rpc error: code = NotFound desc = could not find container \"5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65\": container with ID starting with 5dc6c930b5de26f812087318d22c8bd0eef2406153ea62926485eaa9120cca65 not found: ID does not exist" Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.914436 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7944cbccd6-fnqbk"] Oct 10 10:07:32 crc kubenswrapper[4669]: I1010 10:07:32.924672 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7944cbccd6-fnqbk"] Oct 10 10:07:33 crc kubenswrapper[4669]: I1010 10:07:33.809366 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0a1408-8b73-4070-b491-74d331bcf567" path="/var/lib/kubelet/pods/af0a1408-8b73-4070-b491-74d331bcf567/volumes" Oct 10 10:07:37 crc kubenswrapper[4669]: I1010 10:07:37.238144 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 10 10:07:41 crc kubenswrapper[4669]: I1010 10:07:41.756390 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 10:07:43 crc kubenswrapper[4669]: I1010 10:07:43.755873 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.275397 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.276141 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.276213 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.277510 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1f7630f1b3eab7f0d934e2b66664e0f5dce72b53d2fc00baa4932a36d925d92"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.277863 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://b1f7630f1b3eab7f0d934e2b66664e0f5dce72b53d2fc00baa4932a36d925d92" gracePeriod=600 Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.774334 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="b1f7630f1b3eab7f0d934e2b66664e0f5dce72b53d2fc00baa4932a36d925d92" exitCode=0 Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.774419 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"b1f7630f1b3eab7f0d934e2b66664e0f5dce72b53d2fc00baa4932a36d925d92"} Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.774795 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691"} Oct 10 10:07:54 crc kubenswrapper[4669]: I1010 10:07:54.774824 4669 scope.go:117] "RemoveContainer" containerID="1d2164c0c60f191d64b15377910a1e2b206d868d73ab75e84d53ffcf9dd44ce3" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.595715 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nf6s4"] Oct 10 10:08:27 crc kubenswrapper[4669]: E1010 10:08:27.598278 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon-log" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.598305 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon-log" Oct 10 10:08:27 crc kubenswrapper[4669]: E1010 10:08:27.598527 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.598540 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.601178 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.601257 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0a1408-8b73-4070-b491-74d331bcf567" containerName="horizon-log" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.605985 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.611942 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf6s4"] Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.708045 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5jq2\" (UniqueName: \"kubernetes.io/projected/87863463-157d-4b8a-94b8-98359bba162d-kube-api-access-w5jq2\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.708123 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-utilities\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.708372 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-catalog-content\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.812530 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-catalog-content\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.813841 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-catalog-content\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.814329 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5jq2\" (UniqueName: \"kubernetes.io/projected/87863463-157d-4b8a-94b8-98359bba162d-kube-api-access-w5jq2\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.814517 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-utilities\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.815847 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-utilities\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.835265 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5jq2\" (UniqueName: \"kubernetes.io/projected/87863463-157d-4b8a-94b8-98359bba162d-kube-api-access-w5jq2\") pod \"redhat-marketplace-nf6s4\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:27 crc kubenswrapper[4669]: I1010 10:08:27.948820 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:28 crc kubenswrapper[4669]: I1010 10:08:28.458407 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf6s4"] Oct 10 10:08:29 crc kubenswrapper[4669]: I1010 10:08:29.140432 4669 generic.go:334] "Generic (PLEG): container finished" podID="87863463-157d-4b8a-94b8-98359bba162d" containerID="edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402" exitCode=0 Oct 10 10:08:29 crc kubenswrapper[4669]: I1010 10:08:29.140490 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf6s4" event={"ID":"87863463-157d-4b8a-94b8-98359bba162d","Type":"ContainerDied","Data":"edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402"} Oct 10 10:08:29 crc kubenswrapper[4669]: I1010 10:08:29.140758 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf6s4" event={"ID":"87863463-157d-4b8a-94b8-98359bba162d","Type":"ContainerStarted","Data":"5554ad6697e29da128646a3f1fc698c33d0e8b9b43a1cbf84a93837f8dc6a217"} Oct 10 10:08:30 crc kubenswrapper[4669]: I1010 10:08:30.153180 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf6s4" event={"ID":"87863463-157d-4b8a-94b8-98359bba162d","Type":"ContainerStarted","Data":"ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661"} Oct 10 10:08:31 crc kubenswrapper[4669]: I1010 10:08:31.164421 4669 generic.go:334] "Generic (PLEG): container finished" podID="87863463-157d-4b8a-94b8-98359bba162d" containerID="ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661" exitCode=0 Oct 10 10:08:31 crc kubenswrapper[4669]: I1010 10:08:31.164529 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf6s4" event={"ID":"87863463-157d-4b8a-94b8-98359bba162d","Type":"ContainerDied","Data":"ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661"} Oct 10 10:08:32 crc kubenswrapper[4669]: I1010 10:08:32.177665 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf6s4" event={"ID":"87863463-157d-4b8a-94b8-98359bba162d","Type":"ContainerStarted","Data":"623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a"} Oct 10 10:08:32 crc kubenswrapper[4669]: I1010 10:08:32.209417 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nf6s4" podStartSLOduration=2.536070466 podStartE2EDuration="5.209390604s" podCreationTimestamp="2025-10-10 10:08:27 +0000 UTC" firstStartedPulling="2025-10-10 10:08:29.143627445 +0000 UTC m=+3452.159646227" lastFinishedPulling="2025-10-10 10:08:31.816947623 +0000 UTC m=+3454.832966365" observedRunningTime="2025-10-10 10:08:32.199215978 +0000 UTC m=+3455.215234740" watchObservedRunningTime="2025-10-10 10:08:32.209390604 +0000 UTC m=+3455.225409386" Oct 10 10:08:37 crc kubenswrapper[4669]: I1010 10:08:37.949548 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:37 crc kubenswrapper[4669]: I1010 10:08:37.950257 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:38 crc kubenswrapper[4669]: I1010 10:08:38.000188 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:38 crc kubenswrapper[4669]: I1010 10:08:38.334282 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:38 crc kubenswrapper[4669]: I1010 10:08:38.396515 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf6s4"] Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.260344 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nf6s4" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="registry-server" containerID="cri-o://623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a" gracePeriod=2 Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.775972 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.917029 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5jq2\" (UniqueName: \"kubernetes.io/projected/87863463-157d-4b8a-94b8-98359bba162d-kube-api-access-w5jq2\") pod \"87863463-157d-4b8a-94b8-98359bba162d\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.917937 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-utilities\") pod \"87863463-157d-4b8a-94b8-98359bba162d\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.918133 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-catalog-content\") pod \"87863463-157d-4b8a-94b8-98359bba162d\" (UID: \"87863463-157d-4b8a-94b8-98359bba162d\") " Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.918666 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-utilities" (OuterVolumeSpecName: "utilities") pod "87863463-157d-4b8a-94b8-98359bba162d" (UID: "87863463-157d-4b8a-94b8-98359bba162d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.931675 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87863463-157d-4b8a-94b8-98359bba162d" (UID: "87863463-157d-4b8a-94b8-98359bba162d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.932078 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87863463-157d-4b8a-94b8-98359bba162d-kube-api-access-w5jq2" (OuterVolumeSpecName: "kube-api-access-w5jq2") pod "87863463-157d-4b8a-94b8-98359bba162d" (UID: "87863463-157d-4b8a-94b8-98359bba162d"). InnerVolumeSpecName "kube-api-access-w5jq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.935707 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5jq2\" (UniqueName: \"kubernetes.io/projected/87863463-157d-4b8a-94b8-98359bba162d-kube-api-access-w5jq2\") on node \"crc\" DevicePath \"\"" Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.935730 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:08:40 crc kubenswrapper[4669]: I1010 10:08:40.935757 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87863463-157d-4b8a-94b8-98359bba162d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.275053 4669 generic.go:334] "Generic (PLEG): container finished" podID="87863463-157d-4b8a-94b8-98359bba162d" containerID="623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a" exitCode=0 Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.275095 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf6s4" event={"ID":"87863463-157d-4b8a-94b8-98359bba162d","Type":"ContainerDied","Data":"623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a"} Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.275128 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nf6s4" event={"ID":"87863463-157d-4b8a-94b8-98359bba162d","Type":"ContainerDied","Data":"5554ad6697e29da128646a3f1fc698c33d0e8b9b43a1cbf84a93837f8dc6a217"} Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.275150 4669 scope.go:117] "RemoveContainer" containerID="623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.275207 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nf6s4" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.318203 4669 scope.go:117] "RemoveContainer" containerID="ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.324400 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf6s4"] Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.343761 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nf6s4"] Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.344521 4669 scope.go:117] "RemoveContainer" containerID="edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.385274 4669 scope.go:117] "RemoveContainer" containerID="623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a" Oct 10 10:08:41 crc kubenswrapper[4669]: E1010 10:08:41.386283 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a\": container with ID starting with 623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a not found: ID does not exist" containerID="623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.386330 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a"} err="failed to get container status \"623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a\": rpc error: code = NotFound desc = could not find container \"623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a\": container with ID starting with 623bbaec29187bfc505d845f2a85f52acdb07a546a217e94a27539c99490545a not found: ID does not exist" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.386361 4669 scope.go:117] "RemoveContainer" containerID="ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661" Oct 10 10:08:41 crc kubenswrapper[4669]: E1010 10:08:41.386800 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661\": container with ID starting with ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661 not found: ID does not exist" containerID="ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.386843 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661"} err="failed to get container status \"ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661\": rpc error: code = NotFound desc = could not find container \"ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661\": container with ID starting with ced3d6ea1691c06ab201bfb0d6a1a8ab2f307eee3dbd0acc688968f511bd5661 not found: ID does not exist" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.386869 4669 scope.go:117] "RemoveContainer" containerID="edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402" Oct 10 10:08:41 crc kubenswrapper[4669]: E1010 10:08:41.387254 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402\": container with ID starting with edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402 not found: ID does not exist" containerID="edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.387325 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402"} err="failed to get container status \"edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402\": rpc error: code = NotFound desc = could not find container \"edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402\": container with ID starting with edb0b3cc6de16cc81e13ca40507851431f052793c78c9bc6f6fbf72d90d99402 not found: ID does not exist" Oct 10 10:08:41 crc kubenswrapper[4669]: I1010 10:08:41.809299 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87863463-157d-4b8a-94b8-98359bba162d" path="/var/lib/kubelet/pods/87863463-157d-4b8a-94b8-98359bba162d/volumes" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.848563 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 10 10:08:52 crc kubenswrapper[4669]: E1010 10:08:52.849616 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="extract-utilities" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.849634 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="extract-utilities" Oct 10 10:08:52 crc kubenswrapper[4669]: E1010 10:08:52.849687 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="registry-server" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.849696 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="registry-server" Oct 10 10:08:52 crc kubenswrapper[4669]: E1010 10:08:52.849709 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="extract-content" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.849716 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="extract-content" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.849932 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="87863463-157d-4b8a-94b8-98359bba162d" containerName="registry-server" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.850752 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.859233 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.859472 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5j7sh" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.859654 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.859866 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 10 10:08:52 crc kubenswrapper[4669]: I1010 10:08:52.863144 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.003894 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.003956 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7ckv\" (UniqueName: \"kubernetes.io/projected/732227c2-1b04-4af1-9c97-98e7dd2c4760-kube-api-access-z7ckv\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.003988 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.004185 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.004409 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.004573 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-config-data\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.004721 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.004891 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.005023 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107041 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7ckv\" (UniqueName: \"kubernetes.io/projected/732227c2-1b04-4af1-9c97-98e7dd2c4760-kube-api-access-z7ckv\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107101 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107219 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107280 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107332 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-config-data\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107373 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107492 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107562 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.107625 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.110385 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.110938 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-config-data\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.111287 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.111426 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.111903 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.115335 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.117306 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.121123 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.123336 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7ckv\" (UniqueName: \"kubernetes.io/projected/732227c2-1b04-4af1-9c97-98e7dd2c4760-kube-api-access-z7ckv\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.145940 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.192833 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 10:08:53 crc kubenswrapper[4669]: I1010 10:08:53.667628 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 10 10:08:54 crc kubenswrapper[4669]: I1010 10:08:54.419654 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"732227c2-1b04-4af1-9c97-98e7dd2c4760","Type":"ContainerStarted","Data":"e0acfeaec84c691d11e97f1f784ee5fd28f9fe2981c98014946ce71ee820c1df"} Oct 10 10:09:24 crc kubenswrapper[4669]: E1010 10:09:24.164305 4669 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 10 10:09:24 crc kubenswrapper[4669]: E1010 10:09:24.167554 4669 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z7ckv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(732227c2-1b04-4af1-9c97-98e7dd2c4760): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 10:09:24 crc kubenswrapper[4669]: E1010 10:09:24.169203 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="732227c2-1b04-4af1-9c97-98e7dd2c4760" Oct 10 10:09:24 crc kubenswrapper[4669]: E1010 10:09:24.740799 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="732227c2-1b04-4af1-9c97-98e7dd2c4760" Oct 10 10:09:38 crc kubenswrapper[4669]: I1010 10:09:38.290713 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 10 10:09:39 crc kubenswrapper[4669]: I1010 10:09:39.882263 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"732227c2-1b04-4af1-9c97-98e7dd2c4760","Type":"ContainerStarted","Data":"91fbb8c808f70626061c371931f86d97bcd335a25f84dcc08b22bce6b1b6d282"} Oct 10 10:09:39 crc kubenswrapper[4669]: I1010 10:09:39.910389 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.299043261 podStartE2EDuration="48.910367891s" podCreationTimestamp="2025-10-10 10:08:51 +0000 UTC" firstStartedPulling="2025-10-10 10:08:53.675527125 +0000 UTC m=+3476.691545877" lastFinishedPulling="2025-10-10 10:09:38.286851765 +0000 UTC m=+3521.302870507" observedRunningTime="2025-10-10 10:09:39.907671145 +0000 UTC m=+3522.923689877" watchObservedRunningTime="2025-10-10 10:09:39.910367891 +0000 UTC m=+3522.926386623" Oct 10 10:09:54 crc kubenswrapper[4669]: I1010 10:09:54.275739 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:09:54 crc kubenswrapper[4669]: I1010 10:09:54.276461 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:10:24 crc kubenswrapper[4669]: I1010 10:10:24.275176 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:10:24 crc kubenswrapper[4669]: I1010 10:10:24.277140 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.274944 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.275886 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.277672 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.278659 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.278746 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" gracePeriod=600 Oct 10 10:10:54 crc kubenswrapper[4669]: E1010 10:10:54.416789 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.706138 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" exitCode=0 Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.706211 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691"} Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.706548 4669 scope.go:117] "RemoveContainer" containerID="b1f7630f1b3eab7f0d934e2b66664e0f5dce72b53d2fc00baa4932a36d925d92" Oct 10 10:10:54 crc kubenswrapper[4669]: I1010 10:10:54.707822 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:10:54 crc kubenswrapper[4669]: E1010 10:10:54.708240 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:11:09 crc kubenswrapper[4669]: I1010 10:11:09.796081 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:11:09 crc kubenswrapper[4669]: E1010 10:11:09.796818 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:11:20 crc kubenswrapper[4669]: I1010 10:11:20.795713 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:11:20 crc kubenswrapper[4669]: E1010 10:11:20.797969 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:11:28 crc kubenswrapper[4669]: I1010 10:11:28.890774 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gkmdw"] Oct 10 10:11:28 crc kubenswrapper[4669]: I1010 10:11:28.893776 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:28 crc kubenswrapper[4669]: I1010 10:11:28.924287 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gkmdw"] Oct 10 10:11:28 crc kubenswrapper[4669]: I1010 10:11:28.980150 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-catalog-content\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:28 crc kubenswrapper[4669]: I1010 10:11:28.980299 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chjxr\" (UniqueName: \"kubernetes.io/projected/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-kube-api-access-chjxr\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:28 crc kubenswrapper[4669]: I1010 10:11:28.980410 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-utilities\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.082740 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-catalog-content\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.083140 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chjxr\" (UniqueName: \"kubernetes.io/projected/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-kube-api-access-chjxr\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.083186 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-utilities\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.083372 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-catalog-content\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.083776 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-utilities\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.106570 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chjxr\" (UniqueName: \"kubernetes.io/projected/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-kube-api-access-chjxr\") pod \"certified-operators-gkmdw\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.220126 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:29 crc kubenswrapper[4669]: I1010 10:11:29.737362 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gkmdw"] Oct 10 10:11:30 crc kubenswrapper[4669]: I1010 10:11:30.080961 4669 generic.go:334] "Generic (PLEG): container finished" podID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerID="0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f" exitCode=0 Oct 10 10:11:30 crc kubenswrapper[4669]: I1010 10:11:30.081044 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkmdw" event={"ID":"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1","Type":"ContainerDied","Data":"0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f"} Oct 10 10:11:30 crc kubenswrapper[4669]: I1010 10:11:30.081292 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkmdw" event={"ID":"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1","Type":"ContainerStarted","Data":"8e466e1d68ca21cfcd9eedd0c17f7d16ac71074f7e64199e5e7a210bf32e4d9c"} Oct 10 10:11:30 crc kubenswrapper[4669]: I1010 10:11:30.082717 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 10:11:31 crc kubenswrapper[4669]: I1010 10:11:31.091708 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkmdw" event={"ID":"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1","Type":"ContainerStarted","Data":"a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693"} Oct 10 10:11:33 crc kubenswrapper[4669]: I1010 10:11:33.112260 4669 generic.go:334] "Generic (PLEG): container finished" podID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerID="a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693" exitCode=0 Oct 10 10:11:33 crc kubenswrapper[4669]: I1010 10:11:33.112480 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkmdw" event={"ID":"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1","Type":"ContainerDied","Data":"a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693"} Oct 10 10:11:34 crc kubenswrapper[4669]: I1010 10:11:34.128130 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkmdw" event={"ID":"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1","Type":"ContainerStarted","Data":"0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d"} Oct 10 10:11:34 crc kubenswrapper[4669]: I1010 10:11:34.797263 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:11:34 crc kubenswrapper[4669]: E1010 10:11:34.797873 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:11:39 crc kubenswrapper[4669]: I1010 10:11:39.221001 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:39 crc kubenswrapper[4669]: I1010 10:11:39.221556 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:40 crc kubenswrapper[4669]: I1010 10:11:40.268193 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-gkmdw" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="registry-server" probeResult="failure" output=< Oct 10 10:11:40 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 10:11:40 crc kubenswrapper[4669]: > Oct 10 10:11:47 crc kubenswrapper[4669]: I1010 10:11:47.812277 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:11:47 crc kubenswrapper[4669]: E1010 10:11:47.813764 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:11:49 crc kubenswrapper[4669]: I1010 10:11:49.317952 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:49 crc kubenswrapper[4669]: I1010 10:11:49.342413 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gkmdw" podStartSLOduration=17.839654494 podStartE2EDuration="21.342393094s" podCreationTimestamp="2025-10-10 10:11:28 +0000 UTC" firstStartedPulling="2025-10-10 10:11:30.082482352 +0000 UTC m=+3633.098501084" lastFinishedPulling="2025-10-10 10:11:33.585220942 +0000 UTC m=+3636.601239684" observedRunningTime="2025-10-10 10:11:34.156530835 +0000 UTC m=+3637.172549577" watchObservedRunningTime="2025-10-10 10:11:49.342393094 +0000 UTC m=+3652.358411846" Oct 10 10:11:49 crc kubenswrapper[4669]: I1010 10:11:49.382621 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:49 crc kubenswrapper[4669]: I1010 10:11:49.557571 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gkmdw"] Oct 10 10:11:51 crc kubenswrapper[4669]: I1010 10:11:51.311268 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gkmdw" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="registry-server" containerID="cri-o://0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d" gracePeriod=2 Oct 10 10:11:51 crc kubenswrapper[4669]: I1010 10:11:51.830719 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:51 crc kubenswrapper[4669]: I1010 10:11:51.978257 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chjxr\" (UniqueName: \"kubernetes.io/projected/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-kube-api-access-chjxr\") pod \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " Oct 10 10:11:51 crc kubenswrapper[4669]: I1010 10:11:51.978513 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-catalog-content\") pod \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " Oct 10 10:11:51 crc kubenswrapper[4669]: I1010 10:11:51.978599 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-utilities\") pod \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\" (UID: \"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1\") " Oct 10 10:11:51 crc kubenswrapper[4669]: I1010 10:11:51.980127 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-utilities" (OuterVolumeSpecName: "utilities") pod "98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" (UID: "98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:11:51 crc kubenswrapper[4669]: I1010 10:11:51.987404 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-kube-api-access-chjxr" (OuterVolumeSpecName: "kube-api-access-chjxr") pod "98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" (UID: "98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1"). InnerVolumeSpecName "kube-api-access-chjxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.038907 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" (UID: "98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.083496 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chjxr\" (UniqueName: \"kubernetes.io/projected/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-kube-api-access-chjxr\") on node \"crc\" DevicePath \"\"" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.083577 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.083638 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.323835 4669 generic.go:334] "Generic (PLEG): container finished" podID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerID="0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d" exitCode=0 Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.323898 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkmdw" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.323925 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkmdw" event={"ID":"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1","Type":"ContainerDied","Data":"0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d"} Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.325079 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkmdw" event={"ID":"98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1","Type":"ContainerDied","Data":"8e466e1d68ca21cfcd9eedd0c17f7d16ac71074f7e64199e5e7a210bf32e4d9c"} Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.325115 4669 scope.go:117] "RemoveContainer" containerID="0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.357706 4669 scope.go:117] "RemoveContainer" containerID="a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.370783 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gkmdw"] Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.375540 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gkmdw"] Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.390363 4669 scope.go:117] "RemoveContainer" containerID="0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.422702 4669 scope.go:117] "RemoveContainer" containerID="0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d" Oct 10 10:11:52 crc kubenswrapper[4669]: E1010 10:11:52.423266 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d\": container with ID starting with 0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d not found: ID does not exist" containerID="0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.423308 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d"} err="failed to get container status \"0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d\": rpc error: code = NotFound desc = could not find container \"0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d\": container with ID starting with 0ed6be9b16dc2ad30ebdea23c250cc72bd0d39c9f2a99ad4b199ef5044bf442d not found: ID does not exist" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.423336 4669 scope.go:117] "RemoveContainer" containerID="a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693" Oct 10 10:11:52 crc kubenswrapper[4669]: E1010 10:11:52.423797 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693\": container with ID starting with a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693 not found: ID does not exist" containerID="a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.423827 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693"} err="failed to get container status \"a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693\": rpc error: code = NotFound desc = could not find container \"a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693\": container with ID starting with a8ad6ef8db62c4e8340e01867d13022ef0d068cf8843e4ff154171b05a8ca693 not found: ID does not exist" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.423846 4669 scope.go:117] "RemoveContainer" containerID="0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f" Oct 10 10:11:52 crc kubenswrapper[4669]: E1010 10:11:52.424151 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f\": container with ID starting with 0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f not found: ID does not exist" containerID="0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f" Oct 10 10:11:52 crc kubenswrapper[4669]: I1010 10:11:52.424200 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f"} err="failed to get container status \"0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f\": rpc error: code = NotFound desc = could not find container \"0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f\": container with ID starting with 0d5dc5be26c66733b020138a5a8f186d3fc57903aea2c6cd569edc9132dc304f not found: ID does not exist" Oct 10 10:11:53 crc kubenswrapper[4669]: I1010 10:11:53.814567 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" path="/var/lib/kubelet/pods/98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1/volumes" Oct 10 10:11:58 crc kubenswrapper[4669]: I1010 10:11:58.796468 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:11:58 crc kubenswrapper[4669]: E1010 10:11:58.797152 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:12:13 crc kubenswrapper[4669]: I1010 10:12:13.797457 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:12:13 crc kubenswrapper[4669]: E1010 10:12:13.798792 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:12:25 crc kubenswrapper[4669]: I1010 10:12:25.796607 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:12:25 crc kubenswrapper[4669]: E1010 10:12:25.799041 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:12:37 crc kubenswrapper[4669]: I1010 10:12:37.816473 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:12:37 crc kubenswrapper[4669]: E1010 10:12:37.817540 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:12:49 crc kubenswrapper[4669]: I1010 10:12:49.795709 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:12:49 crc kubenswrapper[4669]: E1010 10:12:49.796753 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:13:00 crc kubenswrapper[4669]: I1010 10:13:00.796137 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:13:00 crc kubenswrapper[4669]: E1010 10:13:00.796998 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:13:14 crc kubenswrapper[4669]: I1010 10:13:14.796512 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:13:14 crc kubenswrapper[4669]: E1010 10:13:14.797666 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:13:24 crc kubenswrapper[4669]: I1010 10:13:24.246174 4669 scope.go:117] "RemoveContainer" containerID="3b353bc7c6be1d4fc253ffa39098bdb38680a6a23eddd771648136cf4a61bb66" Oct 10 10:13:24 crc kubenswrapper[4669]: I1010 10:13:24.280760 4669 scope.go:117] "RemoveContainer" containerID="dfcc9b53bea512a3f30c1206648ec32f338cd5db9d67fe6279f79dfaf1c88fdd" Oct 10 10:13:24 crc kubenswrapper[4669]: I1010 10:13:24.318014 4669 scope.go:117] "RemoveContainer" containerID="9c2014172c974c4e836671fb0ddb6f089a96e4f1f24f587792fde533a05a045f" Oct 10 10:13:24 crc kubenswrapper[4669]: I1010 10:13:24.372674 4669 scope.go:117] "RemoveContainer" containerID="cfe0c805c01223ad752701d12a5bbe96de5c40d1d92ac6125da8422cb7636e53" Oct 10 10:13:29 crc kubenswrapper[4669]: I1010 10:13:29.795650 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:13:29 crc kubenswrapper[4669]: E1010 10:13:29.796685 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.372375 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r5rqt"] Oct 10 10:13:38 crc kubenswrapper[4669]: E1010 10:13:38.375353 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="extract-content" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.375493 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="extract-content" Oct 10 10:13:38 crc kubenswrapper[4669]: E1010 10:13:38.375611 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="registry-server" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.375920 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="registry-server" Oct 10 10:13:38 crc kubenswrapper[4669]: E1010 10:13:38.376034 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="extract-utilities" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.376115 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="extract-utilities" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.376575 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ea4a0d-6e3f-46e3-96a5-f6f5b4f880f1" containerName="registry-server" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.381869 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.381886 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5rqt"] Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.446721 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-catalog-content\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.446797 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhvdj\" (UniqueName: \"kubernetes.io/projected/602adf4a-0f4c-433a-93b7-f835b4978ed8-kube-api-access-dhvdj\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.446919 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-utilities\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.549474 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhvdj\" (UniqueName: \"kubernetes.io/projected/602adf4a-0f4c-433a-93b7-f835b4978ed8-kube-api-access-dhvdj\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.549548 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-catalog-content\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.549646 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-utilities\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.550168 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-catalog-content\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.550236 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-utilities\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.570657 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhvdj\" (UniqueName: \"kubernetes.io/projected/602adf4a-0f4c-433a-93b7-f835b4978ed8-kube-api-access-dhvdj\") pod \"redhat-operators-r5rqt\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:38 crc kubenswrapper[4669]: I1010 10:13:38.713027 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:39 crc kubenswrapper[4669]: I1010 10:13:39.200776 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5rqt"] Oct 10 10:13:39 crc kubenswrapper[4669]: I1010 10:13:39.493160 4669 generic.go:334] "Generic (PLEG): container finished" podID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerID="9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb" exitCode=0 Oct 10 10:13:39 crc kubenswrapper[4669]: I1010 10:13:39.493275 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rqt" event={"ID":"602adf4a-0f4c-433a-93b7-f835b4978ed8","Type":"ContainerDied","Data":"9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb"} Oct 10 10:13:39 crc kubenswrapper[4669]: I1010 10:13:39.493649 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rqt" event={"ID":"602adf4a-0f4c-433a-93b7-f835b4978ed8","Type":"ContainerStarted","Data":"f66756386c6021e215c683e2e3638d42011a24d86312b01c905ccf3ebfc71824"} Oct 10 10:13:41 crc kubenswrapper[4669]: I1010 10:13:41.513423 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rqt" event={"ID":"602adf4a-0f4c-433a-93b7-f835b4978ed8","Type":"ContainerStarted","Data":"dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112"} Oct 10 10:13:43 crc kubenswrapper[4669]: I1010 10:13:43.796334 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:13:43 crc kubenswrapper[4669]: E1010 10:13:43.798578 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:13:45 crc kubenswrapper[4669]: I1010 10:13:45.551274 4669 generic.go:334] "Generic (PLEG): container finished" podID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerID="dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112" exitCode=0 Oct 10 10:13:45 crc kubenswrapper[4669]: I1010 10:13:45.551330 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rqt" event={"ID":"602adf4a-0f4c-433a-93b7-f835b4978ed8","Type":"ContainerDied","Data":"dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112"} Oct 10 10:13:46 crc kubenswrapper[4669]: I1010 10:13:46.562432 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rqt" event={"ID":"602adf4a-0f4c-433a-93b7-f835b4978ed8","Type":"ContainerStarted","Data":"01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21"} Oct 10 10:13:46 crc kubenswrapper[4669]: I1010 10:13:46.614836 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r5rqt" podStartSLOduration=2.144206374 podStartE2EDuration="8.614807216s" podCreationTimestamp="2025-10-10 10:13:38 +0000 UTC" firstStartedPulling="2025-10-10 10:13:39.494896732 +0000 UTC m=+3762.510915474" lastFinishedPulling="2025-10-10 10:13:45.965497574 +0000 UTC m=+3768.981516316" observedRunningTime="2025-10-10 10:13:46.607391068 +0000 UTC m=+3769.623409850" watchObservedRunningTime="2025-10-10 10:13:46.614807216 +0000 UTC m=+3769.630825968" Oct 10 10:13:48 crc kubenswrapper[4669]: I1010 10:13:48.713802 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:48 crc kubenswrapper[4669]: I1010 10:13:48.714840 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:49 crc kubenswrapper[4669]: I1010 10:13:49.768095 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r5rqt" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="registry-server" probeResult="failure" output=< Oct 10 10:13:49 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 10:13:49 crc kubenswrapper[4669]: > Oct 10 10:13:56 crc kubenswrapper[4669]: I1010 10:13:56.275397 4669 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6d556ffc9f-2x4qt" podUID="5bf9895c-806b-4a83-ae94-7530013432fb" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 10 10:13:57 crc kubenswrapper[4669]: I1010 10:13:57.833065 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:13:57 crc kubenswrapper[4669]: E1010 10:13:57.833877 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:13:58 crc kubenswrapper[4669]: I1010 10:13:58.781271 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:58 crc kubenswrapper[4669]: I1010 10:13:58.843812 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:13:59 crc kubenswrapper[4669]: I1010 10:13:59.025760 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5rqt"] Oct 10 10:14:00 crc kubenswrapper[4669]: I1010 10:14:00.695749 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r5rqt" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="registry-server" containerID="cri-o://01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21" gracePeriod=2 Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.329831 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.389270 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-utilities\") pod \"602adf4a-0f4c-433a-93b7-f835b4978ed8\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.389413 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhvdj\" (UniqueName: \"kubernetes.io/projected/602adf4a-0f4c-433a-93b7-f835b4978ed8-kube-api-access-dhvdj\") pod \"602adf4a-0f4c-433a-93b7-f835b4978ed8\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.389464 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-catalog-content\") pod \"602adf4a-0f4c-433a-93b7-f835b4978ed8\" (UID: \"602adf4a-0f4c-433a-93b7-f835b4978ed8\") " Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.390023 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-utilities" (OuterVolumeSpecName: "utilities") pod "602adf4a-0f4c-433a-93b7-f835b4978ed8" (UID: "602adf4a-0f4c-433a-93b7-f835b4978ed8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.390137 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.412292 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/602adf4a-0f4c-433a-93b7-f835b4978ed8-kube-api-access-dhvdj" (OuterVolumeSpecName: "kube-api-access-dhvdj") pod "602adf4a-0f4c-433a-93b7-f835b4978ed8" (UID: "602adf4a-0f4c-433a-93b7-f835b4978ed8"). InnerVolumeSpecName "kube-api-access-dhvdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.492402 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhvdj\" (UniqueName: \"kubernetes.io/projected/602adf4a-0f4c-433a-93b7-f835b4978ed8-kube-api-access-dhvdj\") on node \"crc\" DevicePath \"\"" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.506757 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "602adf4a-0f4c-433a-93b7-f835b4978ed8" (UID: "602adf4a-0f4c-433a-93b7-f835b4978ed8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.595518 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/602adf4a-0f4c-433a-93b7-f835b4978ed8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.710789 4669 generic.go:334] "Generic (PLEG): container finished" podID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerID="01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21" exitCode=0 Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.712645 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rqt" event={"ID":"602adf4a-0f4c-433a-93b7-f835b4978ed8","Type":"ContainerDied","Data":"01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21"} Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.712839 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5rqt" event={"ID":"602adf4a-0f4c-433a-93b7-f835b4978ed8","Type":"ContainerDied","Data":"f66756386c6021e215c683e2e3638d42011a24d86312b01c905ccf3ebfc71824"} Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.713004 4669 scope.go:117] "RemoveContainer" containerID="01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.713063 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5rqt" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.737988 4669 scope.go:117] "RemoveContainer" containerID="dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.771989 4669 scope.go:117] "RemoveContainer" containerID="9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.775389 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5rqt"] Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.783355 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r5rqt"] Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.813417 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" path="/var/lib/kubelet/pods/602adf4a-0f4c-433a-93b7-f835b4978ed8/volumes" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.839319 4669 scope.go:117] "RemoveContainer" containerID="01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21" Oct 10 10:14:01 crc kubenswrapper[4669]: E1010 10:14:01.839937 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21\": container with ID starting with 01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21 not found: ID does not exist" containerID="01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.839976 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21"} err="failed to get container status \"01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21\": rpc error: code = NotFound desc = could not find container \"01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21\": container with ID starting with 01401717c6638a7da90f091a66116aa2bc6d771b1e8adc473fb342aed5c97f21 not found: ID does not exist" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.840003 4669 scope.go:117] "RemoveContainer" containerID="dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112" Oct 10 10:14:01 crc kubenswrapper[4669]: E1010 10:14:01.840291 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112\": container with ID starting with dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112 not found: ID does not exist" containerID="dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.840316 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112"} err="failed to get container status \"dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112\": rpc error: code = NotFound desc = could not find container \"dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112\": container with ID starting with dcc31d4ced5701e7dfbdfc18740a751d4687123f71adc3e000045b10048ce112 not found: ID does not exist" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.840331 4669 scope.go:117] "RemoveContainer" containerID="9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb" Oct 10 10:14:01 crc kubenswrapper[4669]: E1010 10:14:01.840546 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb\": container with ID starting with 9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb not found: ID does not exist" containerID="9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb" Oct 10 10:14:01 crc kubenswrapper[4669]: I1010 10:14:01.840572 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb"} err="failed to get container status \"9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb\": rpc error: code = NotFound desc = could not find container \"9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb\": container with ID starting with 9397ad513effd2ca1b6982fa03d9262432f899a2af9c5491413b519e6c4bd1bb not found: ID does not exist" Oct 10 10:14:12 crc kubenswrapper[4669]: I1010 10:14:12.795901 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:14:12 crc kubenswrapper[4669]: E1010 10:14:12.796856 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:14:25 crc kubenswrapper[4669]: I1010 10:14:25.795189 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:14:25 crc kubenswrapper[4669]: E1010 10:14:25.796394 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:14:40 crc kubenswrapper[4669]: I1010 10:14:40.796604 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:14:40 crc kubenswrapper[4669]: E1010 10:14:40.798044 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:14:51 crc kubenswrapper[4669]: I1010 10:14:51.795719 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:14:51 crc kubenswrapper[4669]: E1010 10:14:51.796800 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.161562 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf"] Oct 10 10:15:00 crc kubenswrapper[4669]: E1010 10:15:00.163484 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="extract-content" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.163513 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="extract-content" Oct 10 10:15:00 crc kubenswrapper[4669]: E1010 10:15:00.163533 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="registry-server" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.163542 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="registry-server" Oct 10 10:15:00 crc kubenswrapper[4669]: E1010 10:15:00.163570 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="extract-utilities" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.163642 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="extract-utilities" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.163989 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="602adf4a-0f4c-433a-93b7-f835b4978ed8" containerName="registry-server" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.165002 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.169122 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.169683 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.188502 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf"] Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.282748 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwmwk\" (UniqueName: \"kubernetes.io/projected/7b2337ef-5af3-4694-b4e4-11261be9b473-kube-api-access-dwmwk\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.283459 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b2337ef-5af3-4694-b4e4-11261be9b473-secret-volume\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.283752 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b2337ef-5af3-4694-b4e4-11261be9b473-config-volume\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.385730 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b2337ef-5af3-4694-b4e4-11261be9b473-secret-volume\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.385832 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b2337ef-5af3-4694-b4e4-11261be9b473-config-volume\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.385927 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwmwk\" (UniqueName: \"kubernetes.io/projected/7b2337ef-5af3-4694-b4e4-11261be9b473-kube-api-access-dwmwk\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.387309 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b2337ef-5af3-4694-b4e4-11261be9b473-config-volume\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.397267 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b2337ef-5af3-4694-b4e4-11261be9b473-secret-volume\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.410022 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwmwk\" (UniqueName: \"kubernetes.io/projected/7b2337ef-5af3-4694-b4e4-11261be9b473-kube-api-access-dwmwk\") pod \"collect-profiles-29334855-kf8vf\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:00 crc kubenswrapper[4669]: I1010 10:15:00.494951 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:01 crc kubenswrapper[4669]: I1010 10:15:01.052972 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf"] Oct 10 10:15:01 crc kubenswrapper[4669]: I1010 10:15:01.325773 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" event={"ID":"7b2337ef-5af3-4694-b4e4-11261be9b473","Type":"ContainerStarted","Data":"b911508bf82bc054eff9d77a4fb2dd45a2f20add9fc4bca65f70b944e8fd22d6"} Oct 10 10:15:01 crc kubenswrapper[4669]: I1010 10:15:01.326254 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" event={"ID":"7b2337ef-5af3-4694-b4e4-11261be9b473","Type":"ContainerStarted","Data":"d34dd80ee2c313c5e7ab28fb95628ef540ec361595a12efbd2384112387430fa"} Oct 10 10:15:01 crc kubenswrapper[4669]: I1010 10:15:01.352130 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" podStartSLOduration=1.352089283 podStartE2EDuration="1.352089283s" podCreationTimestamp="2025-10-10 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:15:01.341689589 +0000 UTC m=+3844.357708331" watchObservedRunningTime="2025-10-10 10:15:01.352089283 +0000 UTC m=+3844.368108035" Oct 10 10:15:02 crc kubenswrapper[4669]: I1010 10:15:02.344317 4669 generic.go:334] "Generic (PLEG): container finished" podID="7b2337ef-5af3-4694-b4e4-11261be9b473" containerID="b911508bf82bc054eff9d77a4fb2dd45a2f20add9fc4bca65f70b944e8fd22d6" exitCode=0 Oct 10 10:15:02 crc kubenswrapper[4669]: I1010 10:15:02.344482 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" event={"ID":"7b2337ef-5af3-4694-b4e4-11261be9b473","Type":"ContainerDied","Data":"b911508bf82bc054eff9d77a4fb2dd45a2f20add9fc4bca65f70b944e8fd22d6"} Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.808097 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.880037 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b2337ef-5af3-4694-b4e4-11261be9b473-config-volume\") pod \"7b2337ef-5af3-4694-b4e4-11261be9b473\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.880254 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b2337ef-5af3-4694-b4e4-11261be9b473-secret-volume\") pod \"7b2337ef-5af3-4694-b4e4-11261be9b473\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.880389 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwmwk\" (UniqueName: \"kubernetes.io/projected/7b2337ef-5af3-4694-b4e4-11261be9b473-kube-api-access-dwmwk\") pod \"7b2337ef-5af3-4694-b4e4-11261be9b473\" (UID: \"7b2337ef-5af3-4694-b4e4-11261be9b473\") " Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.881255 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b2337ef-5af3-4694-b4e4-11261be9b473-config-volume" (OuterVolumeSpecName: "config-volume") pod "7b2337ef-5af3-4694-b4e4-11261be9b473" (UID: "7b2337ef-5af3-4694-b4e4-11261be9b473"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.882374 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b2337ef-5af3-4694-b4e4-11261be9b473-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.893761 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b2337ef-5af3-4694-b4e4-11261be9b473-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7b2337ef-5af3-4694-b4e4-11261be9b473" (UID: "7b2337ef-5af3-4694-b4e4-11261be9b473"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.914152 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b2337ef-5af3-4694-b4e4-11261be9b473-kube-api-access-dwmwk" (OuterVolumeSpecName: "kube-api-access-dwmwk") pod "7b2337ef-5af3-4694-b4e4-11261be9b473" (UID: "7b2337ef-5af3-4694-b4e4-11261be9b473"). InnerVolumeSpecName "kube-api-access-dwmwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.984913 4669 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7b2337ef-5af3-4694-b4e4-11261be9b473-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 10:15:03 crc kubenswrapper[4669]: I1010 10:15:03.984957 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwmwk\" (UniqueName: \"kubernetes.io/projected/7b2337ef-5af3-4694-b4e4-11261be9b473-kube-api-access-dwmwk\") on node \"crc\" DevicePath \"\"" Oct 10 10:15:04 crc kubenswrapper[4669]: I1010 10:15:04.372825 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" event={"ID":"7b2337ef-5af3-4694-b4e4-11261be9b473","Type":"ContainerDied","Data":"d34dd80ee2c313c5e7ab28fb95628ef540ec361595a12efbd2384112387430fa"} Oct 10 10:15:04 crc kubenswrapper[4669]: I1010 10:15:04.372885 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d34dd80ee2c313c5e7ab28fb95628ef540ec361595a12efbd2384112387430fa" Oct 10 10:15:04 crc kubenswrapper[4669]: I1010 10:15:04.372973 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334855-kf8vf" Oct 10 10:15:04 crc kubenswrapper[4669]: I1010 10:15:04.462024 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8"] Oct 10 10:15:04 crc kubenswrapper[4669]: I1010 10:15:04.478120 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334810-947p8"] Oct 10 10:15:04 crc kubenswrapper[4669]: I1010 10:15:04.795920 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:15:04 crc kubenswrapper[4669]: E1010 10:15:04.796976 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:15:05 crc kubenswrapper[4669]: I1010 10:15:05.807695 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00b0f55-ba9f-4cb9-8192-b2db96bc73bc" path="/var/lib/kubelet/pods/a00b0f55-ba9f-4cb9-8192-b2db96bc73bc/volumes" Oct 10 10:15:17 crc kubenswrapper[4669]: I1010 10:15:17.809603 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:15:17 crc kubenswrapper[4669]: E1010 10:15:17.810727 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:15:24 crc kubenswrapper[4669]: I1010 10:15:24.519784 4669 scope.go:117] "RemoveContainer" containerID="01ba41c0373d6f2a0feb7ebeee3024dc63a9151aaa86eec89df7fa5d3f05e8ff" Oct 10 10:15:30 crc kubenswrapper[4669]: I1010 10:15:30.796315 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:15:30 crc kubenswrapper[4669]: E1010 10:15:30.797497 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:15:44 crc kubenswrapper[4669]: I1010 10:15:44.796970 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:15:44 crc kubenswrapper[4669]: E1010 10:15:44.797904 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:15:56 crc kubenswrapper[4669]: I1010 10:15:56.796533 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:15:57 crc kubenswrapper[4669]: I1010 10:15:57.984993 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"21761711a1139aede1cc013920337d2147db34a89ef269bbb87bfc8310a02957"} Oct 10 10:16:09 crc kubenswrapper[4669]: I1010 10:16:09.055180 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-mbmzv"] Oct 10 10:16:09 crc kubenswrapper[4669]: I1010 10:16:09.064736 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-mbmzv"] Oct 10 10:16:09 crc kubenswrapper[4669]: I1010 10:16:09.809698 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbdaa724-fc17-4ef9-9c22-5949776c81f4" path="/var/lib/kubelet/pods/bbdaa724-fc17-4ef9-9c22-5949776c81f4/volumes" Oct 10 10:16:23 crc kubenswrapper[4669]: I1010 10:16:23.039430 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-1e19-account-create-gllwp"] Oct 10 10:16:23 crc kubenswrapper[4669]: I1010 10:16:23.052253 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-1e19-account-create-gllwp"] Oct 10 10:16:23 crc kubenswrapper[4669]: I1010 10:16:23.807270 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fddf3cf-5756-477f-9dd1-c653f747776d" path="/var/lib/kubelet/pods/7fddf3cf-5756-477f-9dd1-c653f747776d/volumes" Oct 10 10:16:24 crc kubenswrapper[4669]: I1010 10:16:24.597223 4669 scope.go:117] "RemoveContainer" containerID="872f3eef90fb9aa1b9705d34e1141ed1e6d6830ba4b7ca0f74a70c61e5838f55" Oct 10 10:16:24 crc kubenswrapper[4669]: I1010 10:16:24.621135 4669 scope.go:117] "RemoveContainer" containerID="4204cf0fbb1bcc7c38f8bf4735936415aaae00b388b794a6f99d487922c8131d" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.015074 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bx4gl"] Oct 10 10:16:28 crc kubenswrapper[4669]: E1010 10:16:28.016405 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b2337ef-5af3-4694-b4e4-11261be9b473" containerName="collect-profiles" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.016421 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b2337ef-5af3-4694-b4e4-11261be9b473" containerName="collect-profiles" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.017323 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b2337ef-5af3-4694-b4e4-11261be9b473" containerName="collect-profiles" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.019085 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.045973 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bx4gl"] Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.129074 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-utilities\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.129217 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcjbs\" (UniqueName: \"kubernetes.io/projected/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-kube-api-access-zcjbs\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.129259 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-catalog-content\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.230746 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcjbs\" (UniqueName: \"kubernetes.io/projected/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-kube-api-access-zcjbs\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.230818 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-catalog-content\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.230877 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-utilities\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.231379 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-utilities\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.232059 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-catalog-content\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.273927 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcjbs\" (UniqueName: \"kubernetes.io/projected/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-kube-api-access-zcjbs\") pod \"community-operators-bx4gl\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.339135 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:28 crc kubenswrapper[4669]: I1010 10:16:28.937803 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bx4gl"] Oct 10 10:16:29 crc kubenswrapper[4669]: I1010 10:16:29.333051 4669 generic.go:334] "Generic (PLEG): container finished" podID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerID="0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3" exitCode=0 Oct 10 10:16:29 crc kubenswrapper[4669]: I1010 10:16:29.333100 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bx4gl" event={"ID":"82e80c0e-5ffa-4488-8ff2-399343d6a1e7","Type":"ContainerDied","Data":"0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3"} Oct 10 10:16:29 crc kubenswrapper[4669]: I1010 10:16:29.333128 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bx4gl" event={"ID":"82e80c0e-5ffa-4488-8ff2-399343d6a1e7","Type":"ContainerStarted","Data":"3447843bf0f81846d024212a7f49203c0ef0869da37ac1b82de583addb113193"} Oct 10 10:16:30 crc kubenswrapper[4669]: I1010 10:16:30.349815 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bx4gl" event={"ID":"82e80c0e-5ffa-4488-8ff2-399343d6a1e7","Type":"ContainerStarted","Data":"43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad"} Oct 10 10:16:32 crc kubenswrapper[4669]: I1010 10:16:32.370703 4669 generic.go:334] "Generic (PLEG): container finished" podID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerID="43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad" exitCode=0 Oct 10 10:16:32 crc kubenswrapper[4669]: I1010 10:16:32.370759 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bx4gl" event={"ID":"82e80c0e-5ffa-4488-8ff2-399343d6a1e7","Type":"ContainerDied","Data":"43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad"} Oct 10 10:16:32 crc kubenswrapper[4669]: I1010 10:16:32.374413 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 10:16:34 crc kubenswrapper[4669]: I1010 10:16:34.392702 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bx4gl" event={"ID":"82e80c0e-5ffa-4488-8ff2-399343d6a1e7","Type":"ContainerStarted","Data":"533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d"} Oct 10 10:16:34 crc kubenswrapper[4669]: I1010 10:16:34.425345 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bx4gl" podStartSLOduration=2.913447617 podStartE2EDuration="7.42532256s" podCreationTimestamp="2025-10-10 10:16:27 +0000 UTC" firstStartedPulling="2025-10-10 10:16:29.336399599 +0000 UTC m=+3932.352418341" lastFinishedPulling="2025-10-10 10:16:33.848274522 +0000 UTC m=+3936.864293284" observedRunningTime="2025-10-10 10:16:34.41937566 +0000 UTC m=+3937.435394402" watchObservedRunningTime="2025-10-10 10:16:34.42532256 +0000 UTC m=+3937.441341312" Oct 10 10:16:38 crc kubenswrapper[4669]: I1010 10:16:38.340987 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:38 crc kubenswrapper[4669]: I1010 10:16:38.341923 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:38 crc kubenswrapper[4669]: I1010 10:16:38.410364 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:47 crc kubenswrapper[4669]: I1010 10:16:47.048040 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-b47bn"] Oct 10 10:16:47 crc kubenswrapper[4669]: I1010 10:16:47.058126 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-b47bn"] Oct 10 10:16:47 crc kubenswrapper[4669]: I1010 10:16:47.807605 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f753198c-0e37-4047-915c-3511c37c70bc" path="/var/lib/kubelet/pods/f753198c-0e37-4047-915c-3511c37c70bc/volumes" Oct 10 10:16:48 crc kubenswrapper[4669]: I1010 10:16:48.391441 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:48 crc kubenswrapper[4669]: I1010 10:16:48.453125 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bx4gl"] Oct 10 10:16:48 crc kubenswrapper[4669]: I1010 10:16:48.522188 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bx4gl" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="registry-server" containerID="cri-o://533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d" gracePeriod=2 Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.024171 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.180433 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-utilities\") pod \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.180621 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-catalog-content\") pod \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.180829 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcjbs\" (UniqueName: \"kubernetes.io/projected/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-kube-api-access-zcjbs\") pod \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\" (UID: \"82e80c0e-5ffa-4488-8ff2-399343d6a1e7\") " Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.181743 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-utilities" (OuterVolumeSpecName: "utilities") pod "82e80c0e-5ffa-4488-8ff2-399343d6a1e7" (UID: "82e80c0e-5ffa-4488-8ff2-399343d6a1e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.187677 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-kube-api-access-zcjbs" (OuterVolumeSpecName: "kube-api-access-zcjbs") pod "82e80c0e-5ffa-4488-8ff2-399343d6a1e7" (UID: "82e80c0e-5ffa-4488-8ff2-399343d6a1e7"). InnerVolumeSpecName "kube-api-access-zcjbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.254820 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82e80c0e-5ffa-4488-8ff2-399343d6a1e7" (UID: "82e80c0e-5ffa-4488-8ff2-399343d6a1e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.284061 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcjbs\" (UniqueName: \"kubernetes.io/projected/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-kube-api-access-zcjbs\") on node \"crc\" DevicePath \"\"" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.284096 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.284109 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82e80c0e-5ffa-4488-8ff2-399343d6a1e7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.533549 4669 generic.go:334] "Generic (PLEG): container finished" podID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerID="533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d" exitCode=0 Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.533700 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bx4gl" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.533729 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bx4gl" event={"ID":"82e80c0e-5ffa-4488-8ff2-399343d6a1e7","Type":"ContainerDied","Data":"533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d"} Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.535245 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bx4gl" event={"ID":"82e80c0e-5ffa-4488-8ff2-399343d6a1e7","Type":"ContainerDied","Data":"3447843bf0f81846d024212a7f49203c0ef0869da37ac1b82de583addb113193"} Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.535291 4669 scope.go:117] "RemoveContainer" containerID="533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.565742 4669 scope.go:117] "RemoveContainer" containerID="43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.574797 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bx4gl"] Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.591832 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bx4gl"] Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.603904 4669 scope.go:117] "RemoveContainer" containerID="0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.650303 4669 scope.go:117] "RemoveContainer" containerID="533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d" Oct 10 10:16:49 crc kubenswrapper[4669]: E1010 10:16:49.651076 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d\": container with ID starting with 533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d not found: ID does not exist" containerID="533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.651139 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d"} err="failed to get container status \"533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d\": rpc error: code = NotFound desc = could not find container \"533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d\": container with ID starting with 533b78c8c5944b56906174f2da5ca8a3ee3fc80b82ee474a0ab9bc9eb3ebce4d not found: ID does not exist" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.651175 4669 scope.go:117] "RemoveContainer" containerID="43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad" Oct 10 10:16:49 crc kubenswrapper[4669]: E1010 10:16:49.651552 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad\": container with ID starting with 43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad not found: ID does not exist" containerID="43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.651612 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad"} err="failed to get container status \"43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad\": rpc error: code = NotFound desc = could not find container \"43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad\": container with ID starting with 43c044da89cae14e98ef70031e093c3833d1b29240bd624d5bc7e58dca6ff5ad not found: ID does not exist" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.651634 4669 scope.go:117] "RemoveContainer" containerID="0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3" Oct 10 10:16:49 crc kubenswrapper[4669]: E1010 10:16:49.654718 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3\": container with ID starting with 0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3 not found: ID does not exist" containerID="0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.654788 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3"} err="failed to get container status \"0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3\": rpc error: code = NotFound desc = could not find container \"0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3\": container with ID starting with 0eae325d7143be45fc6593903727eae7c06f712458a0fb2ac7b2d877d1389fc3 not found: ID does not exist" Oct 10 10:16:49 crc kubenswrapper[4669]: I1010 10:16:49.817396 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" path="/var/lib/kubelet/pods/82e80c0e-5ffa-4488-8ff2-399343d6a1e7/volumes" Oct 10 10:17:24 crc kubenswrapper[4669]: I1010 10:17:24.750684 4669 scope.go:117] "RemoveContainer" containerID="fb5b226bd80f4b0b3e5d065a74c50163a2d3ace424db322cbf66283f32bf3f7f" Oct 10 10:18:24 crc kubenswrapper[4669]: I1010 10:18:24.275009 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:18:24 crc kubenswrapper[4669]: I1010 10:18:24.275641 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.172494 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mqw8m"] Oct 10 10:18:45 crc kubenswrapper[4669]: E1010 10:18:45.173676 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="registry-server" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.173691 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="registry-server" Oct 10 10:18:45 crc kubenswrapper[4669]: E1010 10:18:45.173711 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="extract-content" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.173717 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="extract-content" Oct 10 10:18:45 crc kubenswrapper[4669]: E1010 10:18:45.173731 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="extract-utilities" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.173738 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="extract-utilities" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.173933 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="82e80c0e-5ffa-4488-8ff2-399343d6a1e7" containerName="registry-server" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.176065 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.199445 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqw8m"] Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.346687 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-utilities\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.347157 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96zzv\" (UniqueName: \"kubernetes.io/projected/0ca0c06a-fc23-42a1-9166-d25c86ff8851-kube-api-access-96zzv\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.347309 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-catalog-content\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.449658 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-utilities\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.449835 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96zzv\" (UniqueName: \"kubernetes.io/projected/0ca0c06a-fc23-42a1-9166-d25c86ff8851-kube-api-access-96zzv\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.449919 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-catalog-content\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.450548 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-catalog-content\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.451030 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-utilities\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.474502 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96zzv\" (UniqueName: \"kubernetes.io/projected/0ca0c06a-fc23-42a1-9166-d25c86ff8851-kube-api-access-96zzv\") pod \"redhat-marketplace-mqw8m\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:45 crc kubenswrapper[4669]: I1010 10:18:45.501493 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:46 crc kubenswrapper[4669]: I1010 10:18:46.101193 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqw8m"] Oct 10 10:18:46 crc kubenswrapper[4669]: I1010 10:18:46.759388 4669 generic.go:334] "Generic (PLEG): container finished" podID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerID="f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811" exitCode=0 Oct 10 10:18:46 crc kubenswrapper[4669]: I1010 10:18:46.761247 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqw8m" event={"ID":"0ca0c06a-fc23-42a1-9166-d25c86ff8851","Type":"ContainerDied","Data":"f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811"} Oct 10 10:18:46 crc kubenswrapper[4669]: I1010 10:18:46.761310 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqw8m" event={"ID":"0ca0c06a-fc23-42a1-9166-d25c86ff8851","Type":"ContainerStarted","Data":"f7927d33e1429d97160adfc82b664ae316e2215c241b13a913526aa894232cd9"} Oct 10 10:18:47 crc kubenswrapper[4669]: I1010 10:18:47.774198 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqw8m" event={"ID":"0ca0c06a-fc23-42a1-9166-d25c86ff8851","Type":"ContainerStarted","Data":"b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd"} Oct 10 10:18:48 crc kubenswrapper[4669]: I1010 10:18:48.786034 4669 generic.go:334] "Generic (PLEG): container finished" podID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerID="b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd" exitCode=0 Oct 10 10:18:48 crc kubenswrapper[4669]: I1010 10:18:48.786152 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqw8m" event={"ID":"0ca0c06a-fc23-42a1-9166-d25c86ff8851","Type":"ContainerDied","Data":"b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd"} Oct 10 10:18:49 crc kubenswrapper[4669]: I1010 10:18:49.805055 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqw8m" event={"ID":"0ca0c06a-fc23-42a1-9166-d25c86ff8851","Type":"ContainerStarted","Data":"c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5"} Oct 10 10:18:49 crc kubenswrapper[4669]: I1010 10:18:49.836076 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mqw8m" podStartSLOduration=2.287578511 podStartE2EDuration="4.836055251s" podCreationTimestamp="2025-10-10 10:18:45 +0000 UTC" firstStartedPulling="2025-10-10 10:18:46.76362684 +0000 UTC m=+4069.779645582" lastFinishedPulling="2025-10-10 10:18:49.31210358 +0000 UTC m=+4072.328122322" observedRunningTime="2025-10-10 10:18:49.832562669 +0000 UTC m=+4072.848581401" watchObservedRunningTime="2025-10-10 10:18:49.836055251 +0000 UTC m=+4072.852073993" Oct 10 10:18:54 crc kubenswrapper[4669]: I1010 10:18:54.274885 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:18:54 crc kubenswrapper[4669]: I1010 10:18:54.275693 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:18:55 crc kubenswrapper[4669]: I1010 10:18:55.501648 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:55 crc kubenswrapper[4669]: I1010 10:18:55.502052 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:55 crc kubenswrapper[4669]: I1010 10:18:55.547985 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:55 crc kubenswrapper[4669]: I1010 10:18:55.947944 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:55 crc kubenswrapper[4669]: I1010 10:18:55.996828 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqw8m"] Oct 10 10:18:57 crc kubenswrapper[4669]: I1010 10:18:57.895122 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mqw8m" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="registry-server" containerID="cri-o://c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5" gracePeriod=2 Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.441825 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.562503 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-utilities\") pod \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.562876 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96zzv\" (UniqueName: \"kubernetes.io/projected/0ca0c06a-fc23-42a1-9166-d25c86ff8851-kube-api-access-96zzv\") pod \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.562957 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-catalog-content\") pod \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\" (UID: \"0ca0c06a-fc23-42a1-9166-d25c86ff8851\") " Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.563529 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-utilities" (OuterVolumeSpecName: "utilities") pod "0ca0c06a-fc23-42a1-9166-d25c86ff8851" (UID: "0ca0c06a-fc23-42a1-9166-d25c86ff8851"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.563765 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.580687 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca0c06a-fc23-42a1-9166-d25c86ff8851-kube-api-access-96zzv" (OuterVolumeSpecName: "kube-api-access-96zzv") pod "0ca0c06a-fc23-42a1-9166-d25c86ff8851" (UID: "0ca0c06a-fc23-42a1-9166-d25c86ff8851"). InnerVolumeSpecName "kube-api-access-96zzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.585852 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ca0c06a-fc23-42a1-9166-d25c86ff8851" (UID: "0ca0c06a-fc23-42a1-9166-d25c86ff8851"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.665653 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96zzv\" (UniqueName: \"kubernetes.io/projected/0ca0c06a-fc23-42a1-9166-d25c86ff8851-kube-api-access-96zzv\") on node \"crc\" DevicePath \"\"" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.665694 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ca0c06a-fc23-42a1-9166-d25c86ff8851-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.908855 4669 generic.go:334] "Generic (PLEG): container finished" podID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerID="c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5" exitCode=0 Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.908913 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqw8m" event={"ID":"0ca0c06a-fc23-42a1-9166-d25c86ff8851","Type":"ContainerDied","Data":"c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5"} Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.908949 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqw8m" event={"ID":"0ca0c06a-fc23-42a1-9166-d25c86ff8851","Type":"ContainerDied","Data":"f7927d33e1429d97160adfc82b664ae316e2215c241b13a913526aa894232cd9"} Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.908984 4669 scope.go:117] "RemoveContainer" containerID="c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.909195 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqw8m" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.941888 4669 scope.go:117] "RemoveContainer" containerID="b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd" Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.948855 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqw8m"] Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.963088 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqw8m"] Oct 10 10:18:58 crc kubenswrapper[4669]: I1010 10:18:58.982453 4669 scope.go:117] "RemoveContainer" containerID="f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811" Oct 10 10:18:59 crc kubenswrapper[4669]: I1010 10:18:59.032493 4669 scope.go:117] "RemoveContainer" containerID="c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5" Oct 10 10:18:59 crc kubenswrapper[4669]: E1010 10:18:59.032963 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5\": container with ID starting with c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5 not found: ID does not exist" containerID="c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5" Oct 10 10:18:59 crc kubenswrapper[4669]: I1010 10:18:59.033007 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5"} err="failed to get container status \"c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5\": rpc error: code = NotFound desc = could not find container \"c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5\": container with ID starting with c2b203bd57a62f8658b6c4da9e858808cbe4eb6eaa9e0f163a29f70e0ba16dd5 not found: ID does not exist" Oct 10 10:18:59 crc kubenswrapper[4669]: I1010 10:18:59.033038 4669 scope.go:117] "RemoveContainer" containerID="b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd" Oct 10 10:18:59 crc kubenswrapper[4669]: E1010 10:18:59.033327 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd\": container with ID starting with b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd not found: ID does not exist" containerID="b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd" Oct 10 10:18:59 crc kubenswrapper[4669]: I1010 10:18:59.033365 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd"} err="failed to get container status \"b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd\": rpc error: code = NotFound desc = could not find container \"b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd\": container with ID starting with b140c7a48680bcb5ccd367d0e5a3e715ce9c33c9f7a817db8d5f20efa65b0cfd not found: ID does not exist" Oct 10 10:18:59 crc kubenswrapper[4669]: I1010 10:18:59.033386 4669 scope.go:117] "RemoveContainer" containerID="f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811" Oct 10 10:18:59 crc kubenswrapper[4669]: E1010 10:18:59.033634 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811\": container with ID starting with f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811 not found: ID does not exist" containerID="f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811" Oct 10 10:18:59 crc kubenswrapper[4669]: I1010 10:18:59.033664 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811"} err="failed to get container status \"f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811\": rpc error: code = NotFound desc = could not find container \"f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811\": container with ID starting with f617d4973eba20d838fa332912ad8aa0fdeacb8c878ff25a45f299af5ccae811 not found: ID does not exist" Oct 10 10:18:59 crc kubenswrapper[4669]: I1010 10:18:59.808008 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" path="/var/lib/kubelet/pods/0ca0c06a-fc23-42a1-9166-d25c86ff8851/volumes" Oct 10 10:19:24 crc kubenswrapper[4669]: I1010 10:19:24.274784 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:19:24 crc kubenswrapper[4669]: I1010 10:19:24.275372 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:19:24 crc kubenswrapper[4669]: I1010 10:19:24.275429 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 10:19:24 crc kubenswrapper[4669]: I1010 10:19:24.276429 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21761711a1139aede1cc013920337d2147db34a89ef269bbb87bfc8310a02957"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 10:19:24 crc kubenswrapper[4669]: I1010 10:19:24.276494 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://21761711a1139aede1cc013920337d2147db34a89ef269bbb87bfc8310a02957" gracePeriod=600 Oct 10 10:19:25 crc kubenswrapper[4669]: I1010 10:19:25.183596 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="21761711a1139aede1cc013920337d2147db34a89ef269bbb87bfc8310a02957" exitCode=0 Oct 10 10:19:25 crc kubenswrapper[4669]: I1010 10:19:25.184201 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"21761711a1139aede1cc013920337d2147db34a89ef269bbb87bfc8310a02957"} Oct 10 10:19:25 crc kubenswrapper[4669]: I1010 10:19:25.184337 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29"} Oct 10 10:19:25 crc kubenswrapper[4669]: I1010 10:19:25.185752 4669 scope.go:117] "RemoveContainer" containerID="b2fb6da99154715d528ed2b71bdbd3ceca9f2abb443a92260d114e7afa863691" Oct 10 10:21:24 crc kubenswrapper[4669]: I1010 10:21:24.275129 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:21:24 crc kubenswrapper[4669]: I1010 10:21:24.275725 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:21:54 crc kubenswrapper[4669]: I1010 10:21:54.275070 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:21:54 crc kubenswrapper[4669]: I1010 10:21:54.275738 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:21:59 crc kubenswrapper[4669]: I1010 10:21:59.727979 4669 generic.go:334] "Generic (PLEG): container finished" podID="732227c2-1b04-4af1-9c97-98e7dd2c4760" containerID="91fbb8c808f70626061c371931f86d97bcd335a25f84dcc08b22bce6b1b6d282" exitCode=0 Oct 10 10:21:59 crc kubenswrapper[4669]: I1010 10:21:59.728071 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"732227c2-1b04-4af1-9c97-98e7dd2c4760","Type":"ContainerDied","Data":"91fbb8c808f70626061c371931f86d97bcd335a25f84dcc08b22bce6b1b6d282"} Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.378286 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513292 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7ckv\" (UniqueName: \"kubernetes.io/projected/732227c2-1b04-4af1-9c97-98e7dd2c4760-kube-api-access-z7ckv\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513356 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-temporary\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513416 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ca-certs\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513481 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513517 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513545 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config-secret\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513564 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-config-data\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513579 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-workdir\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.513737 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ssh-key\") pod \"732227c2-1b04-4af1-9c97-98e7dd2c4760\" (UID: \"732227c2-1b04-4af1-9c97-98e7dd2c4760\") " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.515521 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-config-data" (OuterVolumeSpecName: "config-data") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.515543 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.519797 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.523506 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.523663 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/732227c2-1b04-4af1-9c97-98e7dd2c4760-kube-api-access-z7ckv" (OuterVolumeSpecName: "kube-api-access-z7ckv") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "kube-api-access-z7ckv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.544746 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.550166 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.562781 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.586298 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "732227c2-1b04-4af1-9c97-98e7dd2c4760" (UID: "732227c2-1b04-4af1-9c97-98e7dd2c4760"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617862 4669 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617910 4669 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617926 4669 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617939 4669 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617949 4669 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/732227c2-1b04-4af1-9c97-98e7dd2c4760-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617960 4669 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617968 4669 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/732227c2-1b04-4af1-9c97-98e7dd2c4760-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617977 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7ckv\" (UniqueName: \"kubernetes.io/projected/732227c2-1b04-4af1-9c97-98e7dd2c4760-kube-api-access-z7ckv\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.617986 4669 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/732227c2-1b04-4af1-9c97-98e7dd2c4760-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.649966 4669 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.719748 4669 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.763522 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"732227c2-1b04-4af1-9c97-98e7dd2c4760","Type":"ContainerDied","Data":"e0acfeaec84c691d11e97f1f784ee5fd28f9fe2981c98014946ce71ee820c1df"} Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.763559 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0acfeaec84c691d11e97f1f784ee5fd28f9fe2981c98014946ce71ee820c1df" Oct 10 10:22:01 crc kubenswrapper[4669]: I1010 10:22:01.763649 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.258992 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 10 10:22:09 crc kubenswrapper[4669]: E1010 10:22:09.260119 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="extract-utilities" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.260141 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="extract-utilities" Oct 10 10:22:09 crc kubenswrapper[4669]: E1010 10:22:09.260167 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="extract-content" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.260178 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="extract-content" Oct 10 10:22:09 crc kubenswrapper[4669]: E1010 10:22:09.260199 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="registry-server" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.260211 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="registry-server" Oct 10 10:22:09 crc kubenswrapper[4669]: E1010 10:22:09.260237 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732227c2-1b04-4af1-9c97-98e7dd2c4760" containerName="tempest-tests-tempest-tests-runner" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.260248 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="732227c2-1b04-4af1-9c97-98e7dd2c4760" containerName="tempest-tests-tempest-tests-runner" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.260539 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="732227c2-1b04-4af1-9c97-98e7dd2c4760" containerName="tempest-tests-tempest-tests-runner" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.260568 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca0c06a-fc23-42a1-9166-d25c86ff8851" containerName="registry-server" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.261616 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.263837 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5j7sh" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.271107 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.369034 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk7ql\" (UniqueName: \"kubernetes.io/projected/acf0f799-f812-4c43-8fc1-f018e701eebe-kube-api-access-gk7ql\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"acf0f799-f812-4c43-8fc1-f018e701eebe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.369479 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"acf0f799-f812-4c43-8fc1-f018e701eebe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.472024 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk7ql\" (UniqueName: \"kubernetes.io/projected/acf0f799-f812-4c43-8fc1-f018e701eebe-kube-api-access-gk7ql\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"acf0f799-f812-4c43-8fc1-f018e701eebe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.472094 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"acf0f799-f812-4c43-8fc1-f018e701eebe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.472511 4669 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"acf0f799-f812-4c43-8fc1-f018e701eebe\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.502102 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk7ql\" (UniqueName: \"kubernetes.io/projected/acf0f799-f812-4c43-8fc1-f018e701eebe-kube-api-access-gk7ql\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"acf0f799-f812-4c43-8fc1-f018e701eebe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.505243 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"acf0f799-f812-4c43-8fc1-f018e701eebe\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:09 crc kubenswrapper[4669]: I1010 10:22:09.590733 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 10 10:22:10 crc kubenswrapper[4669]: I1010 10:22:10.043266 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 10 10:22:10 crc kubenswrapper[4669]: I1010 10:22:10.064317 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 10:22:10 crc kubenswrapper[4669]: I1010 10:22:10.882369 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"acf0f799-f812-4c43-8fc1-f018e701eebe","Type":"ContainerStarted","Data":"1e762d770f2f27e11f6c6f448e14edc1660e0fb58b69d0fc585072d6fbfd2a52"} Oct 10 10:22:11 crc kubenswrapper[4669]: I1010 10:22:11.893100 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"acf0f799-f812-4c43-8fc1-f018e701eebe","Type":"ContainerStarted","Data":"46915e92005376b9a072ac744da851005607f9550834e13284e7bfe50c307276"} Oct 10 10:22:11 crc kubenswrapper[4669]: I1010 10:22:11.908251 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.8230264520000001 podStartE2EDuration="2.908226557s" podCreationTimestamp="2025-10-10 10:22:09 +0000 UTC" firstStartedPulling="2025-10-10 10:22:10.064083445 +0000 UTC m=+4273.080102187" lastFinishedPulling="2025-10-10 10:22:11.14928355 +0000 UTC m=+4274.165302292" observedRunningTime="2025-10-10 10:22:11.907807203 +0000 UTC m=+4274.923825945" watchObservedRunningTime="2025-10-10 10:22:11.908226557 +0000 UTC m=+4274.924245299" Oct 10 10:22:24 crc kubenswrapper[4669]: I1010 10:22:24.275221 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:22:24 crc kubenswrapper[4669]: I1010 10:22:24.275931 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:22:24 crc kubenswrapper[4669]: I1010 10:22:24.275988 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 10:22:24 crc kubenswrapper[4669]: I1010 10:22:24.276790 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 10:22:24 crc kubenswrapper[4669]: I1010 10:22:24.276855 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" gracePeriod=600 Oct 10 10:22:24 crc kubenswrapper[4669]: E1010 10:22:24.415455 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:22:25 crc kubenswrapper[4669]: I1010 10:22:25.022903 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" exitCode=0 Oct 10 10:22:25 crc kubenswrapper[4669]: I1010 10:22:25.023004 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29"} Oct 10 10:22:25 crc kubenswrapper[4669]: I1010 10:22:25.023447 4669 scope.go:117] "RemoveContainer" containerID="21761711a1139aede1cc013920337d2147db34a89ef269bbb87bfc8310a02957" Oct 10 10:22:25 crc kubenswrapper[4669]: I1010 10:22:25.025179 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:22:25 crc kubenswrapper[4669]: E1010 10:22:25.026555 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.306751 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kpmlf/must-gather-55m2k"] Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.309947 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.312019 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kpmlf"/"openshift-service-ca.crt" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.313032 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kpmlf"/"kube-root-ca.crt" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.313170 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-kpmlf"/"default-dockercfg-b75rq" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.327791 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kpmlf/must-gather-55m2k"] Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.420074 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llff8\" (UniqueName: \"kubernetes.io/projected/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-kube-api-access-llff8\") pod \"must-gather-55m2k\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.420418 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-must-gather-output\") pod \"must-gather-55m2k\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.522794 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llff8\" (UniqueName: \"kubernetes.io/projected/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-kube-api-access-llff8\") pod \"must-gather-55m2k\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.523108 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-must-gather-output\") pod \"must-gather-55m2k\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.523767 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-must-gather-output\") pod \"must-gather-55m2k\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.546876 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llff8\" (UniqueName: \"kubernetes.io/projected/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-kube-api-access-llff8\") pod \"must-gather-55m2k\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:30 crc kubenswrapper[4669]: I1010 10:22:30.643051 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:22:31 crc kubenswrapper[4669]: I1010 10:22:31.154926 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kpmlf/must-gather-55m2k"] Oct 10 10:22:32 crc kubenswrapper[4669]: I1010 10:22:32.100940 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/must-gather-55m2k" event={"ID":"046356c0-e374-4a1b-99bd-ff92eb5ba8e9","Type":"ContainerStarted","Data":"ff69548e291a32b7cc4fa51b9e657e8c867a241badb2a5953148cf4eea537797"} Oct 10 10:22:36 crc kubenswrapper[4669]: I1010 10:22:36.158491 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/must-gather-55m2k" event={"ID":"046356c0-e374-4a1b-99bd-ff92eb5ba8e9","Type":"ContainerStarted","Data":"82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0"} Oct 10 10:22:37 crc kubenswrapper[4669]: I1010 10:22:37.173319 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/must-gather-55m2k" event={"ID":"046356c0-e374-4a1b-99bd-ff92eb5ba8e9","Type":"ContainerStarted","Data":"5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5"} Oct 10 10:22:37 crc kubenswrapper[4669]: I1010 10:22:37.196751 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kpmlf/must-gather-55m2k" podStartSLOduration=2.637225931 podStartE2EDuration="7.196732976s" podCreationTimestamp="2025-10-10 10:22:30 +0000 UTC" firstStartedPulling="2025-10-10 10:22:31.15907401 +0000 UTC m=+4294.175092752" lastFinishedPulling="2025-10-10 10:22:35.718581055 +0000 UTC m=+4298.734599797" observedRunningTime="2025-10-10 10:22:37.192042056 +0000 UTC m=+4300.208060808" watchObservedRunningTime="2025-10-10 10:22:37.196732976 +0000 UTC m=+4300.212751718" Oct 10 10:22:38 crc kubenswrapper[4669]: I1010 10:22:38.894636 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dv5kt"] Oct 10 10:22:38 crc kubenswrapper[4669]: I1010 10:22:38.918032 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dv5kt"] Oct 10 10:22:38 crc kubenswrapper[4669]: I1010 10:22:38.918166 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.086303 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fnm4\" (UniqueName: \"kubernetes.io/projected/bca394f2-4386-4f3b-94c4-5838cfe0adb1-kube-api-access-4fnm4\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.086406 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-utilities\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.086434 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-catalog-content\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.188112 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fnm4\" (UniqueName: \"kubernetes.io/projected/bca394f2-4386-4f3b-94c4-5838cfe0adb1-kube-api-access-4fnm4\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.188198 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-utilities\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.188228 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-catalog-content\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.188636 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-catalog-content\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.188841 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-utilities\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.238793 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fnm4\" (UniqueName: \"kubernetes.io/projected/bca394f2-4386-4f3b-94c4-5838cfe0adb1-kube-api-access-4fnm4\") pod \"certified-operators-dv5kt\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.535786 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.797100 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:22:39 crc kubenswrapper[4669]: E1010 10:22:39.797440 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:22:39 crc kubenswrapper[4669]: W1010 10:22:39.950970 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbca394f2_4386_4f3b_94c4_5838cfe0adb1.slice/crio-29aa438799dd236744465d27abb02b4ded0990a89f243399cb5bcdc78c19e4e9 WatchSource:0}: Error finding container 29aa438799dd236744465d27abb02b4ded0990a89f243399cb5bcdc78c19e4e9: Status 404 returned error can't find the container with id 29aa438799dd236744465d27abb02b4ded0990a89f243399cb5bcdc78c19e4e9 Oct 10 10:22:39 crc kubenswrapper[4669]: I1010 10:22:39.963837 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dv5kt"] Oct 10 10:22:40 crc kubenswrapper[4669]: I1010 10:22:40.203142 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kt" event={"ID":"bca394f2-4386-4f3b-94c4-5838cfe0adb1","Type":"ContainerStarted","Data":"29aa438799dd236744465d27abb02b4ded0990a89f243399cb5bcdc78c19e4e9"} Oct 10 10:22:41 crc kubenswrapper[4669]: E1010 10:22:41.132892 4669 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.53:37178->38.102.83.53:45881: read tcp 38.102.83.53:37178->38.102.83.53:45881: read: connection reset by peer Oct 10 10:22:41 crc kubenswrapper[4669]: I1010 10:22:41.223694 4669 generic.go:334] "Generic (PLEG): container finished" podID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerID="06abffc046635615b145398c8bbfef401ec64d5feaac20db5fdfc87e863b308d" exitCode=0 Oct 10 10:22:41 crc kubenswrapper[4669]: I1010 10:22:41.223752 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kt" event={"ID":"bca394f2-4386-4f3b-94c4-5838cfe0adb1","Type":"ContainerDied","Data":"06abffc046635615b145398c8bbfef401ec64d5feaac20db5fdfc87e863b308d"} Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.249708 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kt" event={"ID":"bca394f2-4386-4f3b-94c4-5838cfe0adb1","Type":"ContainerStarted","Data":"29a0f7a365b539f85ea9eef04e69365e8723154a4dd0bf989c305a48ec0522b8"} Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.653256 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-tnvds"] Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.654623 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.787577 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5bc69013-d069-4f88-875b-0bcea17bf522-host\") pod \"crc-debug-tnvds\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.787726 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg69b\" (UniqueName: \"kubernetes.io/projected/5bc69013-d069-4f88-875b-0bcea17bf522-kube-api-access-sg69b\") pod \"crc-debug-tnvds\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.890531 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5bc69013-d069-4f88-875b-0bcea17bf522-host\") pod \"crc-debug-tnvds\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.890698 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg69b\" (UniqueName: \"kubernetes.io/projected/5bc69013-d069-4f88-875b-0bcea17bf522-kube-api-access-sg69b\") pod \"crc-debug-tnvds\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:42 crc kubenswrapper[4669]: I1010 10:22:42.891360 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5bc69013-d069-4f88-875b-0bcea17bf522-host\") pod \"crc-debug-tnvds\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:43 crc kubenswrapper[4669]: I1010 10:22:43.061248 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg69b\" (UniqueName: \"kubernetes.io/projected/5bc69013-d069-4f88-875b-0bcea17bf522-kube-api-access-sg69b\") pod \"crc-debug-tnvds\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:43 crc kubenswrapper[4669]: I1010 10:22:43.275573 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:22:44 crc kubenswrapper[4669]: I1010 10:22:44.288693 4669 generic.go:334] "Generic (PLEG): container finished" podID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerID="29a0f7a365b539f85ea9eef04e69365e8723154a4dd0bf989c305a48ec0522b8" exitCode=0 Oct 10 10:22:44 crc kubenswrapper[4669]: I1010 10:22:44.289041 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kt" event={"ID":"bca394f2-4386-4f3b-94c4-5838cfe0adb1","Type":"ContainerDied","Data":"29a0f7a365b539f85ea9eef04e69365e8723154a4dd0bf989c305a48ec0522b8"} Oct 10 10:22:44 crc kubenswrapper[4669]: I1010 10:22:44.300010 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" event={"ID":"5bc69013-d069-4f88-875b-0bcea17bf522","Type":"ContainerStarted","Data":"21b59affceb097072e5226ecc4c933755fb013d427307d2bc012c9ae0f667bbf"} Oct 10 10:22:46 crc kubenswrapper[4669]: I1010 10:22:46.333891 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kt" event={"ID":"bca394f2-4386-4f3b-94c4-5838cfe0adb1","Type":"ContainerStarted","Data":"3fd073d6dc63beca1854626648005c4569156583f0981e7a74d55a3b16dec560"} Oct 10 10:22:46 crc kubenswrapper[4669]: I1010 10:22:46.356467 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dv5kt" podStartSLOduration=4.440948252 podStartE2EDuration="8.356446842s" podCreationTimestamp="2025-10-10 10:22:38 +0000 UTC" firstStartedPulling="2025-10-10 10:22:41.225719312 +0000 UTC m=+4304.241738054" lastFinishedPulling="2025-10-10 10:22:45.141217902 +0000 UTC m=+4308.157236644" observedRunningTime="2025-10-10 10:22:46.353643003 +0000 UTC m=+4309.369661745" watchObservedRunningTime="2025-10-10 10:22:46.356446842 +0000 UTC m=+4309.372465594" Oct 10 10:22:49 crc kubenswrapper[4669]: I1010 10:22:49.537477 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:49 crc kubenswrapper[4669]: I1010 10:22:49.538265 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:50 crc kubenswrapper[4669]: I1010 10:22:50.601686 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dv5kt" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="registry-server" probeResult="failure" output=< Oct 10 10:22:50 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 10:22:50 crc kubenswrapper[4669]: > Oct 10 10:22:50 crc kubenswrapper[4669]: I1010 10:22:50.796702 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:22:50 crc kubenswrapper[4669]: E1010 10:22:50.797084 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:22:55 crc kubenswrapper[4669]: I1010 10:22:55.443311 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" event={"ID":"5bc69013-d069-4f88-875b-0bcea17bf522","Type":"ContainerStarted","Data":"4997f388c59ee08bd25bd04f8a4c41387712ca04cf7a544e85136c9bfa7e42b7"} Oct 10 10:22:55 crc kubenswrapper[4669]: I1010 10:22:55.461903 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" podStartSLOduration=1.6472051730000001 podStartE2EDuration="13.461886531s" podCreationTimestamp="2025-10-10 10:22:42 +0000 UTC" firstStartedPulling="2025-10-10 10:22:43.328720294 +0000 UTC m=+4306.344739036" lastFinishedPulling="2025-10-10 10:22:55.143401652 +0000 UTC m=+4318.159420394" observedRunningTime="2025-10-10 10:22:55.46025713 +0000 UTC m=+4318.476275872" watchObservedRunningTime="2025-10-10 10:22:55.461886531 +0000 UTC m=+4318.477905273" Oct 10 10:22:59 crc kubenswrapper[4669]: I1010 10:22:59.589113 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:59 crc kubenswrapper[4669]: I1010 10:22:59.651349 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:22:59 crc kubenswrapper[4669]: I1010 10:22:59.822859 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dv5kt"] Oct 10 10:23:01 crc kubenswrapper[4669]: I1010 10:23:01.491672 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dv5kt" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="registry-server" containerID="cri-o://3fd073d6dc63beca1854626648005c4569156583f0981e7a74d55a3b16dec560" gracePeriod=2 Oct 10 10:23:02 crc kubenswrapper[4669]: I1010 10:23:02.501738 4669 generic.go:334] "Generic (PLEG): container finished" podID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerID="3fd073d6dc63beca1854626648005c4569156583f0981e7a74d55a3b16dec560" exitCode=0 Oct 10 10:23:02 crc kubenswrapper[4669]: I1010 10:23:02.501808 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kt" event={"ID":"bca394f2-4386-4f3b-94c4-5838cfe0adb1","Type":"ContainerDied","Data":"3fd073d6dc63beca1854626648005c4569156583f0981e7a74d55a3b16dec560"} Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.109783 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.230819 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fnm4\" (UniqueName: \"kubernetes.io/projected/bca394f2-4386-4f3b-94c4-5838cfe0adb1-kube-api-access-4fnm4\") pod \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.230987 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-utilities\") pod \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.231166 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-catalog-content\") pod \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\" (UID: \"bca394f2-4386-4f3b-94c4-5838cfe0adb1\") " Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.231474 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-utilities" (OuterVolumeSpecName: "utilities") pod "bca394f2-4386-4f3b-94c4-5838cfe0adb1" (UID: "bca394f2-4386-4f3b-94c4-5838cfe0adb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.249882 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bca394f2-4386-4f3b-94c4-5838cfe0adb1-kube-api-access-4fnm4" (OuterVolumeSpecName: "kube-api-access-4fnm4") pod "bca394f2-4386-4f3b-94c4-5838cfe0adb1" (UID: "bca394f2-4386-4f3b-94c4-5838cfe0adb1"). InnerVolumeSpecName "kube-api-access-4fnm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.262168 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fnm4\" (UniqueName: \"kubernetes.io/projected/bca394f2-4386-4f3b-94c4-5838cfe0adb1-kube-api-access-4fnm4\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.262195 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.279292 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bca394f2-4386-4f3b-94c4-5838cfe0adb1" (UID: "bca394f2-4386-4f3b-94c4-5838cfe0adb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.366195 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca394f2-4386-4f3b-94c4-5838cfe0adb1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.514668 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kt" event={"ID":"bca394f2-4386-4f3b-94c4-5838cfe0adb1","Type":"ContainerDied","Data":"29aa438799dd236744465d27abb02b4ded0990a89f243399cb5bcdc78c19e4e9"} Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.514712 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kt" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.515920 4669 scope.go:117] "RemoveContainer" containerID="3fd073d6dc63beca1854626648005c4569156583f0981e7a74d55a3b16dec560" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.549757 4669 scope.go:117] "RemoveContainer" containerID="29a0f7a365b539f85ea9eef04e69365e8723154a4dd0bf989c305a48ec0522b8" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.558504 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dv5kt"] Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.573059 4669 scope.go:117] "RemoveContainer" containerID="06abffc046635615b145398c8bbfef401ec64d5feaac20db5fdfc87e863b308d" Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.575827 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dv5kt"] Oct 10 10:23:03 crc kubenswrapper[4669]: I1010 10:23:03.804875 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" path="/var/lib/kubelet/pods/bca394f2-4386-4f3b-94c4-5838cfe0adb1/volumes" Oct 10 10:23:05 crc kubenswrapper[4669]: I1010 10:23:05.796712 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:23:05 crc kubenswrapper[4669]: E1010 10:23:05.797441 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:23:18 crc kubenswrapper[4669]: I1010 10:23:18.796766 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:23:18 crc kubenswrapper[4669]: E1010 10:23:18.798260 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:23:30 crc kubenswrapper[4669]: I1010 10:23:30.795551 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:23:30 crc kubenswrapper[4669]: E1010 10:23:30.796371 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:23:38 crc kubenswrapper[4669]: I1010 10:23:38.874865 4669 generic.go:334] "Generic (PLEG): container finished" podID="5bc69013-d069-4f88-875b-0bcea17bf522" containerID="4997f388c59ee08bd25bd04f8a4c41387712ca04cf7a544e85136c9bfa7e42b7" exitCode=0 Oct 10 10:23:38 crc kubenswrapper[4669]: I1010 10:23:38.876328 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" event={"ID":"5bc69013-d069-4f88-875b-0bcea17bf522","Type":"ContainerDied","Data":"4997f388c59ee08bd25bd04f8a4c41387712ca04cf7a544e85136c9bfa7e42b7"} Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.007667 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.067217 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-tnvds"] Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.080959 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-tnvds"] Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.119128 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5bc69013-d069-4f88-875b-0bcea17bf522-host\") pod \"5bc69013-d069-4f88-875b-0bcea17bf522\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.119359 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg69b\" (UniqueName: \"kubernetes.io/projected/5bc69013-d069-4f88-875b-0bcea17bf522-kube-api-access-sg69b\") pod \"5bc69013-d069-4f88-875b-0bcea17bf522\" (UID: \"5bc69013-d069-4f88-875b-0bcea17bf522\") " Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.119741 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bc69013-d069-4f88-875b-0bcea17bf522-host" (OuterVolumeSpecName: "host") pod "5bc69013-d069-4f88-875b-0bcea17bf522" (UID: "5bc69013-d069-4f88-875b-0bcea17bf522"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.119874 4669 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5bc69013-d069-4f88-875b-0bcea17bf522-host\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.148861 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc69013-d069-4f88-875b-0bcea17bf522-kube-api-access-sg69b" (OuterVolumeSpecName: "kube-api-access-sg69b") pod "5bc69013-d069-4f88-875b-0bcea17bf522" (UID: "5bc69013-d069-4f88-875b-0bcea17bf522"). InnerVolumeSpecName "kube-api-access-sg69b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.222505 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg69b\" (UniqueName: \"kubernetes.io/projected/5bc69013-d069-4f88-875b-0bcea17bf522-kube-api-access-sg69b\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.893964 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21b59affceb097072e5226ecc4c933755fb013d427307d2bc012c9ae0f667bbf" Oct 10 10:23:40 crc kubenswrapper[4669]: I1010 10:23:40.894066 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-tnvds" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.409056 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-mcj5v"] Oct 10 10:23:41 crc kubenswrapper[4669]: E1010 10:23:41.409466 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="registry-server" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.409484 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="registry-server" Oct 10 10:23:41 crc kubenswrapper[4669]: E1010 10:23:41.409501 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc69013-d069-4f88-875b-0bcea17bf522" containerName="container-00" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.409508 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc69013-d069-4f88-875b-0bcea17bf522" containerName="container-00" Oct 10 10:23:41 crc kubenswrapper[4669]: E1010 10:23:41.409538 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="extract-content" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.409547 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="extract-content" Oct 10 10:23:41 crc kubenswrapper[4669]: E1010 10:23:41.409563 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="extract-utilities" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.409570 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="extract-utilities" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.409809 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc69013-d069-4f88-875b-0bcea17bf522" containerName="container-00" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.409838 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="bca394f2-4386-4f3b-94c4-5838cfe0adb1" containerName="registry-server" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.410492 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.547968 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/deea4c66-daec-4666-ba5e-0ccb0da59aad-host\") pod \"crc-debug-mcj5v\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.548031 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfhnj\" (UniqueName: \"kubernetes.io/projected/deea4c66-daec-4666-ba5e-0ccb0da59aad-kube-api-access-pfhnj\") pod \"crc-debug-mcj5v\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.649786 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfhnj\" (UniqueName: \"kubernetes.io/projected/deea4c66-daec-4666-ba5e-0ccb0da59aad-kube-api-access-pfhnj\") pod \"crc-debug-mcj5v\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.649993 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/deea4c66-daec-4666-ba5e-0ccb0da59aad-host\") pod \"crc-debug-mcj5v\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.650087 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/deea4c66-daec-4666-ba5e-0ccb0da59aad-host\") pod \"crc-debug-mcj5v\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.675060 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfhnj\" (UniqueName: \"kubernetes.io/projected/deea4c66-daec-4666-ba5e-0ccb0da59aad-kube-api-access-pfhnj\") pod \"crc-debug-mcj5v\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.733477 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.801324 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:23:41 crc kubenswrapper[4669]: E1010 10:23:41.807487 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.807722 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc69013-d069-4f88-875b-0bcea17bf522" path="/var/lib/kubelet/pods/5bc69013-d069-4f88-875b-0bcea17bf522/volumes" Oct 10 10:23:41 crc kubenswrapper[4669]: I1010 10:23:41.902703 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" event={"ID":"deea4c66-daec-4666-ba5e-0ccb0da59aad","Type":"ContainerStarted","Data":"d564b423371873a1f477104e0059ebe8355411be71b95436c69871c11f8b0b45"} Oct 10 10:23:42 crc kubenswrapper[4669]: I1010 10:23:42.913902 4669 generic.go:334] "Generic (PLEG): container finished" podID="deea4c66-daec-4666-ba5e-0ccb0da59aad" containerID="f9770f8bd8105878b2deee9c942b2946214729d6be47b44a24d21ad12c23314c" exitCode=0 Oct 10 10:23:42 crc kubenswrapper[4669]: I1010 10:23:42.914001 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" event={"ID":"deea4c66-daec-4666-ba5e-0ccb0da59aad","Type":"ContainerDied","Data":"f9770f8bd8105878b2deee9c942b2946214729d6be47b44a24d21ad12c23314c"} Oct 10 10:23:43 crc kubenswrapper[4669]: I1010 10:23:43.350504 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-mcj5v"] Oct 10 10:23:43 crc kubenswrapper[4669]: I1010 10:23:43.358199 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-mcj5v"] Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.029841 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.203431 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfhnj\" (UniqueName: \"kubernetes.io/projected/deea4c66-daec-4666-ba5e-0ccb0da59aad-kube-api-access-pfhnj\") pod \"deea4c66-daec-4666-ba5e-0ccb0da59aad\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.203532 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/deea4c66-daec-4666-ba5e-0ccb0da59aad-host\") pod \"deea4c66-daec-4666-ba5e-0ccb0da59aad\" (UID: \"deea4c66-daec-4666-ba5e-0ccb0da59aad\") " Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.203959 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deea4c66-daec-4666-ba5e-0ccb0da59aad-host" (OuterVolumeSpecName: "host") pod "deea4c66-daec-4666-ba5e-0ccb0da59aad" (UID: "deea4c66-daec-4666-ba5e-0ccb0da59aad"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.204132 4669 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/deea4c66-daec-4666-ba5e-0ccb0da59aad-host\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.212843 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deea4c66-daec-4666-ba5e-0ccb0da59aad-kube-api-access-pfhnj" (OuterVolumeSpecName: "kube-api-access-pfhnj") pod "deea4c66-daec-4666-ba5e-0ccb0da59aad" (UID: "deea4c66-daec-4666-ba5e-0ccb0da59aad"). InnerVolumeSpecName "kube-api-access-pfhnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.305990 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfhnj\" (UniqueName: \"kubernetes.io/projected/deea4c66-daec-4666-ba5e-0ccb0da59aad-kube-api-access-pfhnj\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.596328 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-qgfqh"] Oct 10 10:23:44 crc kubenswrapper[4669]: E1010 10:23:44.597102 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deea4c66-daec-4666-ba5e-0ccb0da59aad" containerName="container-00" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.597117 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="deea4c66-daec-4666-ba5e-0ccb0da59aad" containerName="container-00" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.597305 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="deea4c66-daec-4666-ba5e-0ccb0da59aad" containerName="container-00" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.598519 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.713533 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2399ac80-216e-4758-b624-2fe210a2a839-host\") pod \"crc-debug-qgfqh\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.713921 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnvcj\" (UniqueName: \"kubernetes.io/projected/2399ac80-216e-4758-b624-2fe210a2a839-kube-api-access-bnvcj\") pod \"crc-debug-qgfqh\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.815383 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnvcj\" (UniqueName: \"kubernetes.io/projected/2399ac80-216e-4758-b624-2fe210a2a839-kube-api-access-bnvcj\") pod \"crc-debug-qgfqh\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.815514 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2399ac80-216e-4758-b624-2fe210a2a839-host\") pod \"crc-debug-qgfqh\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.815672 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2399ac80-216e-4758-b624-2fe210a2a839-host\") pod \"crc-debug-qgfqh\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.874629 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnvcj\" (UniqueName: \"kubernetes.io/projected/2399ac80-216e-4758-b624-2fe210a2a839-kube-api-access-bnvcj\") pod \"crc-debug-qgfqh\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.920103 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.945644 4669 scope.go:117] "RemoveContainer" containerID="f9770f8bd8105878b2deee9c942b2946214729d6be47b44a24d21ad12c23314c" Oct 10 10:23:44 crc kubenswrapper[4669]: I1010 10:23:44.945679 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-mcj5v" Oct 10 10:23:45 crc kubenswrapper[4669]: I1010 10:23:45.806328 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deea4c66-daec-4666-ba5e-0ccb0da59aad" path="/var/lib/kubelet/pods/deea4c66-daec-4666-ba5e-0ccb0da59aad/volumes" Oct 10 10:23:45 crc kubenswrapper[4669]: I1010 10:23:45.954928 4669 generic.go:334] "Generic (PLEG): container finished" podID="2399ac80-216e-4758-b624-2fe210a2a839" containerID="dc91ed6fe69deb7ff7127044e479773bf161489a5198521dcba3cf176cd28634" exitCode=0 Oct 10 10:23:45 crc kubenswrapper[4669]: I1010 10:23:45.954992 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" event={"ID":"2399ac80-216e-4758-b624-2fe210a2a839","Type":"ContainerDied","Data":"dc91ed6fe69deb7ff7127044e479773bf161489a5198521dcba3cf176cd28634"} Oct 10 10:23:45 crc kubenswrapper[4669]: I1010 10:23:45.955035 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" event={"ID":"2399ac80-216e-4758-b624-2fe210a2a839","Type":"ContainerStarted","Data":"7647f999d9a5287324bdf23d459a2d4c0f4ee9f7aaaa251b2fb25fcb574c1763"} Oct 10 10:23:45 crc kubenswrapper[4669]: I1010 10:23:45.992296 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-qgfqh"] Oct 10 10:23:46 crc kubenswrapper[4669]: I1010 10:23:46.007155 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kpmlf/crc-debug-qgfqh"] Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.075067 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.177427 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnvcj\" (UniqueName: \"kubernetes.io/projected/2399ac80-216e-4758-b624-2fe210a2a839-kube-api-access-bnvcj\") pod \"2399ac80-216e-4758-b624-2fe210a2a839\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.177639 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2399ac80-216e-4758-b624-2fe210a2a839-host\") pod \"2399ac80-216e-4758-b624-2fe210a2a839\" (UID: \"2399ac80-216e-4758-b624-2fe210a2a839\") " Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.177789 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2399ac80-216e-4758-b624-2fe210a2a839-host" (OuterVolumeSpecName: "host") pod "2399ac80-216e-4758-b624-2fe210a2a839" (UID: "2399ac80-216e-4758-b624-2fe210a2a839"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.178275 4669 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2399ac80-216e-4758-b624-2fe210a2a839-host\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.205363 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2399ac80-216e-4758-b624-2fe210a2a839-kube-api-access-bnvcj" (OuterVolumeSpecName: "kube-api-access-bnvcj") pod "2399ac80-216e-4758-b624-2fe210a2a839" (UID: "2399ac80-216e-4758-b624-2fe210a2a839"). InnerVolumeSpecName "kube-api-access-bnvcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.279714 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnvcj\" (UniqueName: \"kubernetes.io/projected/2399ac80-216e-4758-b624-2fe210a2a839-kube-api-access-bnvcj\") on node \"crc\" DevicePath \"\"" Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.808763 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2399ac80-216e-4758-b624-2fe210a2a839" path="/var/lib/kubelet/pods/2399ac80-216e-4758-b624-2fe210a2a839/volumes" Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.974623 4669 scope.go:117] "RemoveContainer" containerID="dc91ed6fe69deb7ff7127044e479773bf161489a5198521dcba3cf176cd28634" Oct 10 10:23:47 crc kubenswrapper[4669]: I1010 10:23:47.974692 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/crc-debug-qgfqh" Oct 10 10:23:54 crc kubenswrapper[4669]: I1010 10:23:54.795331 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:23:54 crc kubenswrapper[4669]: E1010 10:23:54.796279 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:24:05 crc kubenswrapper[4669]: I1010 10:24:05.796253 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:24:05 crc kubenswrapper[4669]: E1010 10:24:05.797200 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:24:17 crc kubenswrapper[4669]: I1010 10:24:17.808029 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:24:17 crc kubenswrapper[4669]: E1010 10:24:17.808857 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:24:24 crc kubenswrapper[4669]: I1010 10:24:24.372241 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8c86b488-lgxwv_9306ad2c-0c67-4a1a-8031-19d4cddce8d0/barbican-api/0.log" Oct 10 10:24:24 crc kubenswrapper[4669]: I1010 10:24:24.411333 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8c86b488-lgxwv_9306ad2c-0c67-4a1a-8031-19d4cddce8d0/barbican-api-log/0.log" Oct 10 10:24:24 crc kubenswrapper[4669]: I1010 10:24:24.624537 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fcbcfb6bb-9hc2p_bb7dc377-170c-4504-a61b-386110e80526/barbican-keystone-listener-log/0.log" Oct 10 10:24:24 crc kubenswrapper[4669]: I1010 10:24:24.651507 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fcbcfb6bb-9hc2p_bb7dc377-170c-4504-a61b-386110e80526/barbican-keystone-listener/0.log" Oct 10 10:24:24 crc kubenswrapper[4669]: I1010 10:24:24.728791 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-686db9dfc7-hg7z4_58f4d796-043f-48f0-a649-99fb297c694a/barbican-worker/0.log" Oct 10 10:24:24 crc kubenswrapper[4669]: I1010 10:24:24.889094 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-686db9dfc7-hg7z4_58f4d796-043f-48f0-a649-99fb297c694a/barbican-worker-log/0.log" Oct 10 10:24:24 crc kubenswrapper[4669]: I1010 10:24:24.985842 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v_66dbecf2-f169-4c6c-b8be-efbab612ace8/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.147457 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/ceilometer-notification-agent/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.196407 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/ceilometer-central-agent/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.236107 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/proxy-httpd/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.375805 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/sg-core/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.448685 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q_c1b45643-fd51-4a37-a298-f9b84b997c40/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.591745 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv_e53d40a4-c456-4991-8cfc-4dc303b5205d/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.712389 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50ffa58f-0d75-4996-b66c-092aacc84c63/cinder-api/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.732985 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50ffa58f-0d75-4996-b66c-092aacc84c63/cinder-api-log/0.log" Oct 10 10:24:25 crc kubenswrapper[4669]: I1010 10:24:25.947366 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c642048f-8f7f-4355-bd06-e862839e25e9/probe/0.log" Oct 10 10:24:26 crc kubenswrapper[4669]: I1010 10:24:26.086126 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c642048f-8f7f-4355-bd06-e862839e25e9/cinder-backup/0.log" Oct 10 10:24:26 crc kubenswrapper[4669]: I1010 10:24:26.716509 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23eadb50-c141-42eb-b24c-2a493eec134e/cinder-scheduler/0.log" Oct 10 10:24:26 crc kubenswrapper[4669]: I1010 10:24:26.742416 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23eadb50-c141-42eb-b24c-2a493eec134e/probe/0.log" Oct 10 10:24:26 crc kubenswrapper[4669]: I1010 10:24:26.765434 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3c92ae00-f31f-4d27-aaca-8d3920a345eb/cinder-volume/0.log" Oct 10 10:24:26 crc kubenswrapper[4669]: I1010 10:24:26.919829 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3c92ae00-f31f-4d27-aaca-8d3920a345eb/probe/0.log" Oct 10 10:24:27 crc kubenswrapper[4669]: I1010 10:24:27.004138 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg_7469da7f-e608-439c-93b1-d60130d75a72/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:27 crc kubenswrapper[4669]: I1010 10:24:27.141987 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-st5n8_2d5855cf-c66a-4daa-9b94-84636ee8936d/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:27 crc kubenswrapper[4669]: I1010 10:24:27.452023 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58d5f4b747-fl954_2dff0743-e2de-497b-8b6e-8b7773e19da5/init/0.log" Oct 10 10:24:27 crc kubenswrapper[4669]: I1010 10:24:27.552554 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58d5f4b747-fl954_2dff0743-e2de-497b-8b6e-8b7773e19da5/init/0.log" Oct 10 10:24:27 crc kubenswrapper[4669]: I1010 10:24:27.704354 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58d5f4b747-fl954_2dff0743-e2de-497b-8b6e-8b7773e19da5/dnsmasq-dns/0.log" Oct 10 10:24:27 crc kubenswrapper[4669]: I1010 10:24:27.748881 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_60e2fea4-35a0-45d6-bd36-e7172b6822fe/glance-httpd/0.log" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.326931 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_60e2fea4-35a0-45d6-bd36-e7172b6822fe/glance-log/0.log" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.399326 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_731fc6e9-0f93-4e9f-bbc0-767b6fada47b/glance-log/0.log" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.436641 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_731fc6e9-0f93-4e9f-bbc0-767b6fada47b/glance-httpd/0.log" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.662291 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-566f97fbdd-plcxg_0040911a-8dff-49e4-99af-aafb84f7639a/horizon/0.log" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.719207 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b_880204d3-a81a-433f-ace4-ec7ff0092b53/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.795588 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:24:28 crc kubenswrapper[4669]: E1010 10:24:28.795968 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.816113 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-566f97fbdd-plcxg_0040911a-8dff-49e4-99af-aafb84f7639a/horizon-log/0.log" Oct 10 10:24:28 crc kubenswrapper[4669]: I1010 10:24:28.992997 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hzlnz_9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.229867 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29334841-rb65h_d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759/keystone-cron/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.243415 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-65b8b49f6-vm5h8_8c9b81a2-843b-4bab-b58a-5e2a1ce2e636/keystone-api/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.364919 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_0330c42a-ef18-46cd-9a95-5622b53328b7/kube-state-metrics/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.531234 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-8kd62_fbaa41c7-694f-4fda-aaf5-481188158475/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.584771 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_53ead5e1-28e3-4677-99e0-e2faa30ae9a5/manila-api-log/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.644068 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_53ead5e1-28e3-4677-99e0-e2faa30ae9a5/manila-api/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.832032 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_8a1791d4-98d8-4eac-9676-e409ac975611/probe/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.836761 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_8a1791d4-98d8-4eac-9676-e409ac975611/manila-scheduler/0.log" Oct 10 10:24:29 crc kubenswrapper[4669]: I1010 10:24:29.979259 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e4865c42-8dbc-4d96-afb9-e65d85f8ec96/manila-share/0.log" Oct 10 10:24:30 crc kubenswrapper[4669]: I1010 10:24:30.161074 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e4865c42-8dbc-4d96-afb9-e65d85f8ec96/probe/0.log" Oct 10 10:24:30 crc kubenswrapper[4669]: I1010 10:24:30.401248 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6d556ffc9f-2x4qt_5bf9895c-806b-4a83-ae94-7530013432fb/neutron-api/0.log" Oct 10 10:24:30 crc kubenswrapper[4669]: I1010 10:24:30.495182 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6d556ffc9f-2x4qt_5bf9895c-806b-4a83-ae94-7530013432fb/neutron-httpd/0.log" Oct 10 10:24:30 crc kubenswrapper[4669]: I1010 10:24:30.617514 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w_f9d4c45c-5b27-4012-a0d0-72a1a2b1843e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:31 crc kubenswrapper[4669]: I1010 10:24:31.173457 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d81e619a-c272-4824-b19f-dea88a428acd/nova-api-log/0.log" Oct 10 10:24:31 crc kubenswrapper[4669]: I1010 10:24:31.238779 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0c654342-12f4-40f3-a4dd-e4e5a76b8b76/nova-cell0-conductor-conductor/0.log" Oct 10 10:24:31 crc kubenswrapper[4669]: I1010 10:24:31.603738 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d81e619a-c272-4824-b19f-dea88a428acd/nova-api-api/0.log" Oct 10 10:24:31 crc kubenswrapper[4669]: I1010 10:24:31.626888 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_9a6e22fc-ac5c-469b-bf5e-3ce1de48968a/nova-cell1-conductor-conductor/0.log" Oct 10 10:24:31 crc kubenswrapper[4669]: I1010 10:24:31.658553 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5d5e59af-66c7-4723-9758-9369ccf46cde/nova-cell1-novncproxy-novncproxy/0.log" Oct 10 10:24:32 crc kubenswrapper[4669]: I1010 10:24:32.003730 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6_4933ae0b-dda6-44e1-a551-730e803bb7d5/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:32 crc kubenswrapper[4669]: I1010 10:24:32.221294 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f6b1ea56-8a03-4b04-ba07-0852d3fb6545/nova-metadata-log/0.log" Oct 10 10:24:32 crc kubenswrapper[4669]: I1010 10:24:32.886656 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_cbfa50a5-b829-4023-b978-d76660fc77b2/nova-scheduler-scheduler/0.log" Oct 10 10:24:32 crc kubenswrapper[4669]: I1010 10:24:32.907693 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_02229a1f-8349-4e0c-b5b5-b56ba246f395/mysql-bootstrap/0.log" Oct 10 10:24:33 crc kubenswrapper[4669]: I1010 10:24:33.134078 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_02229a1f-8349-4e0c-b5b5-b56ba246f395/galera/0.log" Oct 10 10:24:33 crc kubenswrapper[4669]: I1010 10:24:33.146714 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_02229a1f-8349-4e0c-b5b5-b56ba246f395/mysql-bootstrap/0.log" Oct 10 10:24:33 crc kubenswrapper[4669]: I1010 10:24:33.375052 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_79d402e9-8187-4cd5-9df5-9297d8fc4b5a/mysql-bootstrap/0.log" Oct 10 10:24:33 crc kubenswrapper[4669]: I1010 10:24:33.682240 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f6b1ea56-8a03-4b04-ba07-0852d3fb6545/nova-metadata-metadata/0.log" Oct 10 10:24:33 crc kubenswrapper[4669]: I1010 10:24:33.697525 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_79d402e9-8187-4cd5-9df5-9297d8fc4b5a/mysql-bootstrap/0.log" Oct 10 10:24:33 crc kubenswrapper[4669]: I1010 10:24:33.771570 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_79d402e9-8187-4cd5-9df5-9297d8fc4b5a/galera/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.043515 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jrw5q_2b815c31-96c7-47c3-b458-35945b55b326/ovn-controller/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.070707 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2bb8c92a-28c8-4424-94c2-c577f0b6b486/openstackclient/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.264234 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vg62q_1f2fdb85-cfc4-471d-b35e-9388c3406895/openstack-network-exporter/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.528670 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovsdb-server-init/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.720161 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovsdb-server-init/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.720563 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovsdb-server/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.760846 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovs-vswitchd/0.log" Oct 10 10:24:34 crc kubenswrapper[4669]: I1010 10:24:34.987306 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nsc5k_cd3df673-494f-4e69-a140-34e78f12567c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:35 crc kubenswrapper[4669]: I1010 10:24:35.144998 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_028c96a4-38cf-4710-bcb2-a132836d77f1/openstack-network-exporter/0.log" Oct 10 10:24:35 crc kubenswrapper[4669]: I1010 10:24:35.174084 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_028c96a4-38cf-4710-bcb2-a132836d77f1/ovn-northd/0.log" Oct 10 10:24:35 crc kubenswrapper[4669]: I1010 10:24:35.517427 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5321fb79-5d79-4094-b314-8914ed41cfd5/openstack-network-exporter/0.log" Oct 10 10:24:35 crc kubenswrapper[4669]: I1010 10:24:35.518552 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5321fb79-5d79-4094-b314-8914ed41cfd5/ovsdbserver-nb/0.log" Oct 10 10:24:35 crc kubenswrapper[4669]: I1010 10:24:35.669675 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9f30537-9153-4862-97aa-31e4f5e57ecf/openstack-network-exporter/0.log" Oct 10 10:24:35 crc kubenswrapper[4669]: I1010 10:24:35.776646 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9f30537-9153-4862-97aa-31e4f5e57ecf/ovsdbserver-sb/0.log" Oct 10 10:24:36 crc kubenswrapper[4669]: I1010 10:24:36.014558 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f557bc54-vbfbn_d9be36cc-2136-4285-8489-93149d0136c3/placement-api/0.log" Oct 10 10:24:36 crc kubenswrapper[4669]: I1010 10:24:36.154670 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f557bc54-vbfbn_d9be36cc-2136-4285-8489-93149d0136c3/placement-log/0.log" Oct 10 10:24:36 crc kubenswrapper[4669]: I1010 10:24:36.813008 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05416508-674b-44d8-9074-62776915ec92/setup-container/0.log" Oct 10 10:24:37 crc kubenswrapper[4669]: I1010 10:24:37.055376 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05416508-674b-44d8-9074-62776915ec92/setup-container/0.log" Oct 10 10:24:37 crc kubenswrapper[4669]: I1010 10:24:37.133248 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05416508-674b-44d8-9074-62776915ec92/rabbitmq/0.log" Oct 10 10:24:37 crc kubenswrapper[4669]: I1010 10:24:37.219863 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1172c8c-d2a5-4f42-b97a-87265385580b/setup-container/0.log" Oct 10 10:24:37 crc kubenswrapper[4669]: I1010 10:24:37.503856 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1172c8c-d2a5-4f42-b97a-87265385580b/setup-container/0.log" Oct 10 10:24:37 crc kubenswrapper[4669]: I1010 10:24:37.595614 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c_f9938853-891c-4766-82ef-21e7b718dcc2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:37 crc kubenswrapper[4669]: I1010 10:24:37.628779 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1172c8c-d2a5-4f42-b97a-87265385580b/rabbitmq/0.log" Oct 10 10:24:37 crc kubenswrapper[4669]: I1010 10:24:37.976926 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-768lk_483acfce-519d-4164-a1fe-e8786ade906f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.015986 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh_5c0a91b3-706e-47df-8f2b-84733ad00593/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.428696 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_732227c2-1b04-4af1-9c97-98e7dd2c4760/tempest-tests-tempest-tests-runner/0.log" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.469278 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dpnrk_5dc32d38-3a9b-48bb-93c4-76f51d215213/ssh-known-hosts-edpm-deployment/0.log" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.490618 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kflfv"] Oct 10 10:24:38 crc kubenswrapper[4669]: E1010 10:24:38.491301 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2399ac80-216e-4758-b624-2fe210a2a839" containerName="container-00" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.491323 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="2399ac80-216e-4758-b624-2fe210a2a839" containerName="container-00" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.491569 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="2399ac80-216e-4758-b624-2fe210a2a839" containerName="container-00" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.494237 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.503971 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kflfv"] Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.525153 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-catalog-content\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.525204 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtphn\" (UniqueName: \"kubernetes.io/projected/47678fab-de95-411e-b074-ec581bc8abe1-kube-api-access-mtphn\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.525394 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-utilities\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.626859 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-catalog-content\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.626915 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtphn\" (UniqueName: \"kubernetes.io/projected/47678fab-de95-411e-b074-ec581bc8abe1-kube-api-access-mtphn\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.627007 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-utilities\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.627488 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-utilities\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.627761 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-catalog-content\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.674310 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtphn\" (UniqueName: \"kubernetes.io/projected/47678fab-de95-411e-b074-ec581bc8abe1-kube-api-access-mtphn\") pod \"redhat-operators-kflfv\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:38 crc kubenswrapper[4669]: I1010 10:24:38.848762 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:39 crc kubenswrapper[4669]: I1010 10:24:39.102912 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl_4feb106b-763c-422d-a1df-a95b354a7770/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:24:39 crc kubenswrapper[4669]: I1010 10:24:39.122272 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_acf0f799-f812-4c43-8fc1-f018e701eebe/test-operator-logs-container/0.log" Oct 10 10:24:39 crc kubenswrapper[4669]: I1010 10:24:39.515377 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kflfv"] Oct 10 10:24:39 crc kubenswrapper[4669]: I1010 10:24:39.803459 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:24:39 crc kubenswrapper[4669]: E1010 10:24:39.803718 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:24:40 crc kubenswrapper[4669]: I1010 10:24:40.485779 4669 generic.go:334] "Generic (PLEG): container finished" podID="47678fab-de95-411e-b074-ec581bc8abe1" containerID="88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad" exitCode=0 Oct 10 10:24:40 crc kubenswrapper[4669]: I1010 10:24:40.486026 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kflfv" event={"ID":"47678fab-de95-411e-b074-ec581bc8abe1","Type":"ContainerDied","Data":"88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad"} Oct 10 10:24:40 crc kubenswrapper[4669]: I1010 10:24:40.486056 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kflfv" event={"ID":"47678fab-de95-411e-b074-ec581bc8abe1","Type":"ContainerStarted","Data":"d7d2dc9c846ff86a8ed97b3f1a500d265e7e9d5da7ffda7a1dfe698532f3abff"} Oct 10 10:24:41 crc kubenswrapper[4669]: I1010 10:24:41.504901 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kflfv" event={"ID":"47678fab-de95-411e-b074-ec581bc8abe1","Type":"ContainerStarted","Data":"2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de"} Oct 10 10:24:45 crc kubenswrapper[4669]: I1010 10:24:45.547786 4669 generic.go:334] "Generic (PLEG): container finished" podID="47678fab-de95-411e-b074-ec581bc8abe1" containerID="2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de" exitCode=0 Oct 10 10:24:45 crc kubenswrapper[4669]: I1010 10:24:45.548387 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kflfv" event={"ID":"47678fab-de95-411e-b074-ec581bc8abe1","Type":"ContainerDied","Data":"2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de"} Oct 10 10:24:46 crc kubenswrapper[4669]: I1010 10:24:46.556764 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kflfv" event={"ID":"47678fab-de95-411e-b074-ec581bc8abe1","Type":"ContainerStarted","Data":"4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a"} Oct 10 10:24:46 crc kubenswrapper[4669]: I1010 10:24:46.575412 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kflfv" podStartSLOduration=3.080340135 podStartE2EDuration="8.575395265s" podCreationTimestamp="2025-10-10 10:24:38 +0000 UTC" firstStartedPulling="2025-10-10 10:24:40.489387829 +0000 UTC m=+4423.505406571" lastFinishedPulling="2025-10-10 10:24:45.984442959 +0000 UTC m=+4429.000461701" observedRunningTime="2025-10-10 10:24:46.573345329 +0000 UTC m=+4429.589364071" watchObservedRunningTime="2025-10-10 10:24:46.575395265 +0000 UTC m=+4429.591414007" Oct 10 10:24:48 crc kubenswrapper[4669]: I1010 10:24:48.851960 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:48 crc kubenswrapper[4669]: I1010 10:24:48.852366 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:49 crc kubenswrapper[4669]: I1010 10:24:49.918770 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kflfv" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="registry-server" probeResult="failure" output=< Oct 10 10:24:49 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 10:24:49 crc kubenswrapper[4669]: > Oct 10 10:24:52 crc kubenswrapper[4669]: I1010 10:24:52.253751 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9c376674-4cf1-40c0-9e3c-c25e3fefad30/memcached/0.log" Oct 10 10:24:52 crc kubenswrapper[4669]: I1010 10:24:52.795252 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:24:52 crc kubenswrapper[4669]: E1010 10:24:52.795678 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:24:58 crc kubenswrapper[4669]: I1010 10:24:58.902809 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:58 crc kubenswrapper[4669]: I1010 10:24:58.968829 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:24:59 crc kubenswrapper[4669]: I1010 10:24:59.137910 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kflfv"] Oct 10 10:25:00 crc kubenswrapper[4669]: I1010 10:25:00.664356 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kflfv" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="registry-server" containerID="cri-o://4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a" gracePeriod=2 Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.247659 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.445297 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtphn\" (UniqueName: \"kubernetes.io/projected/47678fab-de95-411e-b074-ec581bc8abe1-kube-api-access-mtphn\") pod \"47678fab-de95-411e-b074-ec581bc8abe1\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.445999 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-catalog-content\") pod \"47678fab-de95-411e-b074-ec581bc8abe1\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.446231 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-utilities\") pod \"47678fab-de95-411e-b074-ec581bc8abe1\" (UID: \"47678fab-de95-411e-b074-ec581bc8abe1\") " Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.447192 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-utilities" (OuterVolumeSpecName: "utilities") pod "47678fab-de95-411e-b074-ec581bc8abe1" (UID: "47678fab-de95-411e-b074-ec581bc8abe1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.451924 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47678fab-de95-411e-b074-ec581bc8abe1-kube-api-access-mtphn" (OuterVolumeSpecName: "kube-api-access-mtphn") pod "47678fab-de95-411e-b074-ec581bc8abe1" (UID: "47678fab-de95-411e-b074-ec581bc8abe1"). InnerVolumeSpecName "kube-api-access-mtphn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.533128 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47678fab-de95-411e-b074-ec581bc8abe1" (UID: "47678fab-de95-411e-b074-ec581bc8abe1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.562794 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtphn\" (UniqueName: \"kubernetes.io/projected/47678fab-de95-411e-b074-ec581bc8abe1-kube-api-access-mtphn\") on node \"crc\" DevicePath \"\"" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.562846 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.562871 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47678fab-de95-411e-b074-ec581bc8abe1-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.675792 4669 generic.go:334] "Generic (PLEG): container finished" podID="47678fab-de95-411e-b074-ec581bc8abe1" containerID="4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a" exitCode=0 Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.675841 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kflfv" event={"ID":"47678fab-de95-411e-b074-ec581bc8abe1","Type":"ContainerDied","Data":"4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a"} Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.675875 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kflfv" event={"ID":"47678fab-de95-411e-b074-ec581bc8abe1","Type":"ContainerDied","Data":"d7d2dc9c846ff86a8ed97b3f1a500d265e7e9d5da7ffda7a1dfe698532f3abff"} Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.675896 4669 scope.go:117] "RemoveContainer" containerID="4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.675926 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kflfv" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.710272 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kflfv"] Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.711432 4669 scope.go:117] "RemoveContainer" containerID="2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.718401 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kflfv"] Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.740222 4669 scope.go:117] "RemoveContainer" containerID="88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.780112 4669 scope.go:117] "RemoveContainer" containerID="4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a" Oct 10 10:25:01 crc kubenswrapper[4669]: E1010 10:25:01.780763 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a\": container with ID starting with 4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a not found: ID does not exist" containerID="4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.780803 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a"} err="failed to get container status \"4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a\": rpc error: code = NotFound desc = could not find container \"4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a\": container with ID starting with 4171dada070e3512c8a764d22b1ddd9269cd41b7dcbf4e49886796bff37d430a not found: ID does not exist" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.780833 4669 scope.go:117] "RemoveContainer" containerID="2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de" Oct 10 10:25:01 crc kubenswrapper[4669]: E1010 10:25:01.784958 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de\": container with ID starting with 2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de not found: ID does not exist" containerID="2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.784999 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de"} err="failed to get container status \"2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de\": rpc error: code = NotFound desc = could not find container \"2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de\": container with ID starting with 2c298f045606a4fae9824cb98d1d7ba670c3beb2135b0569216f2d2f1f3579de not found: ID does not exist" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.785030 4669 scope.go:117] "RemoveContainer" containerID="88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad" Oct 10 10:25:01 crc kubenswrapper[4669]: E1010 10:25:01.785351 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad\": container with ID starting with 88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad not found: ID does not exist" containerID="88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.785384 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad"} err="failed to get container status \"88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad\": rpc error: code = NotFound desc = could not find container \"88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad\": container with ID starting with 88fbb19fb146f94136c4e04ad36a4d645a83ed1dd1177bb2ece534168b99d2ad not found: ID does not exist" Oct 10 10:25:01 crc kubenswrapper[4669]: I1010 10:25:01.811138 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47678fab-de95-411e-b074-ec581bc8abe1" path="/var/lib/kubelet/pods/47678fab-de95-411e-b074-ec581bc8abe1/volumes" Oct 10 10:25:04 crc kubenswrapper[4669]: I1010 10:25:04.798018 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:25:04 crc kubenswrapper[4669]: E1010 10:25:04.799253 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:25:11 crc kubenswrapper[4669]: I1010 10:25:11.478927 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-sbc5v_affd00ea-eac5-4c5f-b452-c8bb0e17aba8/kube-rbac-proxy/0.log" Oct 10 10:25:11 crc kubenswrapper[4669]: I1010 10:25:11.521536 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-sbc5v_affd00ea-eac5-4c5f-b452-c8bb0e17aba8/manager/0.log" Oct 10 10:25:11 crc kubenswrapper[4669]: I1010 10:25:11.708048 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-jzps8_7dcc64bf-ac2d-4357-adcb-b6c146462464/kube-rbac-proxy/0.log" Oct 10 10:25:11 crc kubenswrapper[4669]: I1010 10:25:11.811139 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-jzps8_7dcc64bf-ac2d-4357-adcb-b6c146462464/manager/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.102402 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/util/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.125229 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-pn8xd_1a5e08a1-2d18-4382-8129-be1376c40bde/manager/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.139277 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-pn8xd_1a5e08a1-2d18-4382-8129-be1376c40bde/kube-rbac-proxy/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.410624 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/pull/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.491190 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/util/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.551145 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/pull/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.805631 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/pull/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.807644 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/extract/0.log" Oct 10 10:25:12 crc kubenswrapper[4669]: I1010 10:25:12.809632 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/util/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.048289 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-pfp4n_55f7e993-3dd8-4acf-9716-47cef7f73ef8/kube-rbac-proxy/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.079763 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-pdnnc_bd3dd74a-3c58-4785-854e-400c7ba726de/kube-rbac-proxy/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.095743 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-pfp4n_55f7e993-3dd8-4acf-9716-47cef7f73ef8/manager/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.218802 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-pdnnc_bd3dd74a-3c58-4785-854e-400c7ba726de/manager/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.315323 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-96278_909f202d-6571-4743-9314-26cc27e94c77/kube-rbac-proxy/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.370313 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-96278_909f202d-6571-4743-9314-26cc27e94c77/manager/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.511342 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-czldp_270b1e97-6178-44d9-8d9f-ed5e4dc02000/kube-rbac-proxy/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.680469 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-czldp_270b1e97-6178-44d9-8d9f-ed5e4dc02000/manager/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.729727 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-k5nxp_2132b635-c43e-4c4d-9074-37e81b6345f2/kube-rbac-proxy/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.786211 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-k5nxp_2132b635-c43e-4c4d-9074-37e81b6345f2/manager/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.864387 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-6zprx_27b6cee3-b0a3-4116-b9bb-299a81bba403/kube-rbac-proxy/0.log" Oct 10 10:25:13 crc kubenswrapper[4669]: I1010 10:25:13.978048 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-6zprx_27b6cee3-b0a3-4116-b9bb-299a81bba403/manager/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.041372 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-q52hz_6941897d-fd2e-4ff3-ad50-aeba82eb86c2/kube-rbac-proxy/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.150504 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-q52hz_6941897d-fd2e-4ff3-ad50-aeba82eb86c2/manager/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.226833 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-n72xm_82353170-7f06-43e4-966b-20d0d234f0f2/kube-rbac-proxy/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.288122 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-n72xm_82353170-7f06-43e4-966b-20d0d234f0f2/manager/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.454447 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-kwn4g_bdebd762-e639-4e50-820d-cd32885a1341/kube-rbac-proxy/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.586607 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-kwn4g_bdebd762-e639-4e50-820d-cd32885a1341/manager/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.702009 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-pdf8v_48aca3a8-684a-448f-ab17-b9d604d47484/kube-rbac-proxy/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.819328 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-pdf8v_48aca3a8-684a-448f-ab17-b9d604d47484/manager/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.844529 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-wtsqj_43bf605f-c6c8-447b-9dc3-03be42fe7f10/kube-rbac-proxy/0.log" Oct 10 10:25:14 crc kubenswrapper[4669]: I1010 10:25:14.981857 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-wtsqj_43bf605f-c6c8-447b-9dc3-03be42fe7f10/manager/0.log" Oct 10 10:25:15 crc kubenswrapper[4669]: I1010 10:25:15.049976 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv_73105b8d-0527-4510-886b-7203163993fd/kube-rbac-proxy/0.log" Oct 10 10:25:15 crc kubenswrapper[4669]: I1010 10:25:15.109834 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv_73105b8d-0527-4510-886b-7203163993fd/manager/0.log" Oct 10 10:25:15 crc kubenswrapper[4669]: I1010 10:25:15.248999 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5698bb9464-flrhj_5f2708ab-b3ba-4556-9e11-7018b186975e/kube-rbac-proxy/0.log" Oct 10 10:25:15 crc kubenswrapper[4669]: I1010 10:25:15.547877 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-599bffcb5d-2bj6v_d6df673a-6169-4bcc-bf13-98eb9e3a20f9/kube-rbac-proxy/0.log" Oct 10 10:25:15 crc kubenswrapper[4669]: I1010 10:25:15.797285 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:25:15 crc kubenswrapper[4669]: E1010 10:25:15.798282 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:25:15 crc kubenswrapper[4669]: I1010 10:25:15.851019 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-599bffcb5d-2bj6v_d6df673a-6169-4bcc-bf13-98eb9e3a20f9/operator/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.045449 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2tjnm_0e034df9-3c3c-4b15-8d82-2a4179090ed7/registry-server/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.146057 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-r7ctx_29661927-099c-4549-8584-9390881170d5/kube-rbac-proxy/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.286834 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-r7ctx_29661927-099c-4549-8584-9390881170d5/manager/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.344329 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-4b4jb_316d31d2-d9a9-4fd7-a079-4620ad8c5e09/kube-rbac-proxy/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.574985 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5698bb9464-flrhj_5f2708ab-b3ba-4556-9e11-7018b186975e/manager/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.578438 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-4b4jb_316d31d2-d9a9-4fd7-a079-4620ad8c5e09/manager/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.775131 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs_637efcba-83b3-4f3a-93f0-1ce6257987e1/operator/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.856599 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-hjkht_db1620d5-baef-433a-b9e7-07a55004a68f/manager/0.log" Oct 10 10:25:16 crc kubenswrapper[4669]: I1010 10:25:16.870210 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-hjkht_db1620d5-baef-433a-b9e7-07a55004a68f/kube-rbac-proxy/0.log" Oct 10 10:25:17 crc kubenswrapper[4669]: I1010 10:25:17.064923 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-j7fm7_f30a8280-3312-4333-8966-810b8dd3c889/kube-rbac-proxy/0.log" Oct 10 10:25:17 crc kubenswrapper[4669]: I1010 10:25:17.103470 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-b2p7l_684c8731-c863-49d2-a029-d9d69ae01f8d/kube-rbac-proxy/0.log" Oct 10 10:25:17 crc kubenswrapper[4669]: I1010 10:25:17.173025 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-j7fm7_f30a8280-3312-4333-8966-810b8dd3c889/manager/0.log" Oct 10 10:25:17 crc kubenswrapper[4669]: I1010 10:25:17.645971 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-b2p7l_684c8731-c863-49d2-a029-d9d69ae01f8d/manager/0.log" Oct 10 10:25:17 crc kubenswrapper[4669]: I1010 10:25:17.761433 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-cjtjr_5d2f190a-4eae-44e6-a5cb-e15abf2e037d/manager/0.log" Oct 10 10:25:17 crc kubenswrapper[4669]: I1010 10:25:17.791791 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-cjtjr_5d2f190a-4eae-44e6-a5cb-e15abf2e037d/kube-rbac-proxy/0.log" Oct 10 10:25:30 crc kubenswrapper[4669]: I1010 10:25:30.796012 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:25:30 crc kubenswrapper[4669]: E1010 10:25:30.797082 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:25:36 crc kubenswrapper[4669]: I1010 10:25:36.900940 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-f6dtf_f273bc2e-26d7-40fe-9150-fb582fb14618/control-plane-machine-set-operator/0.log" Oct 10 10:25:37 crc kubenswrapper[4669]: I1010 10:25:37.117070 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9lpfb_e57d53b2-5487-4820-aa61-9070eb866eae/kube-rbac-proxy/0.log" Oct 10 10:25:37 crc kubenswrapper[4669]: I1010 10:25:37.122671 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9lpfb_e57d53b2-5487-4820-aa61-9070eb866eae/machine-api-operator/0.log" Oct 10 10:25:45 crc kubenswrapper[4669]: I1010 10:25:45.795727 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:25:45 crc kubenswrapper[4669]: E1010 10:25:45.796320 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:25:51 crc kubenswrapper[4669]: I1010 10:25:51.567101 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-tqwjt_4ab4da17-72b9-45f6-94c3-75905972615a/cert-manager-controller/0.log" Oct 10 10:25:51 crc kubenswrapper[4669]: I1010 10:25:51.667722 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-45l5f_a9cc3175-a61f-44c4-8668-6270126b84b7/cert-manager-cainjector/0.log" Oct 10 10:25:51 crc kubenswrapper[4669]: I1010 10:25:51.866797 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4258m_302f9809-912d-4e36-b0d5-7bdbae70071a/cert-manager-webhook/0.log" Oct 10 10:25:59 crc kubenswrapper[4669]: I1010 10:25:59.796260 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:25:59 crc kubenswrapper[4669]: E1010 10:25:59.797062 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:26:07 crc kubenswrapper[4669]: I1010 10:26:07.532482 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-brznz_5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0/nmstate-console-plugin/0.log" Oct 10 10:26:07 crc kubenswrapper[4669]: I1010 10:26:07.684074 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-v5rbb_27433356-4791-4652-b258-aaf3d6deebef/nmstate-handler/0.log" Oct 10 10:26:07 crc kubenswrapper[4669]: I1010 10:26:07.847109 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-jlckz_35a376a3-b98d-4c9d-b7a8-633759d368ea/kube-rbac-proxy/0.log" Oct 10 10:26:07 crc kubenswrapper[4669]: I1010 10:26:07.888636 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-jlckz_35a376a3-b98d-4c9d-b7a8-633759d368ea/nmstate-metrics/0.log" Oct 10 10:26:08 crc kubenswrapper[4669]: I1010 10:26:08.152287 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-qwxrv_1b2674b7-3907-4f5a-b3b8-cc6945e4f004/nmstate-operator/0.log" Oct 10 10:26:08 crc kubenswrapper[4669]: I1010 10:26:08.182093 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-dm9zp_9ec015c7-dd0b-46e0-a5b9-456bf97c6a49/nmstate-webhook/0.log" Oct 10 10:26:11 crc kubenswrapper[4669]: I1010 10:26:11.798217 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:26:11 crc kubenswrapper[4669]: E1010 10:26:11.799660 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:26:25 crc kubenswrapper[4669]: I1010 10:26:25.795698 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:26:25 crc kubenswrapper[4669]: E1010 10:26:25.796868 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:26:26 crc kubenswrapper[4669]: I1010 10:26:26.922411 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xh4cl_0eeead67-0b6d-48c7-8756-50e1c1eb5f0b/controller/0.log" Oct 10 10:26:26 crc kubenswrapper[4669]: I1010 10:26:26.944687 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xh4cl_0eeead67-0b6d-48c7-8756-50e1c1eb5f0b/kube-rbac-proxy/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.167850 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.415078 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.450669 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.481732 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.514575 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.702618 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.722365 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.780227 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:26:27 crc kubenswrapper[4669]: I1010 10:26:27.813843 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.025113 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.084526 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.101163 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.155927 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/controller/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.367279 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/kube-rbac-proxy/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.418305 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/frr-metrics/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.497102 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/kube-rbac-proxy-frr/0.log" Oct 10 10:26:28 crc kubenswrapper[4669]: I1010 10:26:28.726477 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/reloader/0.log" Oct 10 10:26:29 crc kubenswrapper[4669]: I1010 10:26:29.494206 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-rgcll_8bd73190-2f9a-45df-8e2d-8579238bf4a3/frr-k8s-webhook-server/0.log" Oct 10 10:26:29 crc kubenswrapper[4669]: I1010 10:26:29.560838 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79d96cb9f5-vkfb4_5cda559e-d9fe-4651-b77d-d8c7d38e6a21/manager/0.log" Oct 10 10:26:29 crc kubenswrapper[4669]: I1010 10:26:29.744557 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/frr/0.log" Oct 10 10:26:29 crc kubenswrapper[4669]: I1010 10:26:29.919023 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7dd78c579-pxx42_06827887-c1ea-4839-9ed3-a4c8813d81ac/webhook-server/0.log" Oct 10 10:26:30 crc kubenswrapper[4669]: I1010 10:26:30.066490 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-78hzw_ec1f1180-0070-43bd-9ee3-095e8e2379f0/kube-rbac-proxy/0.log" Oct 10 10:26:30 crc kubenswrapper[4669]: I1010 10:26:30.414469 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-78hzw_ec1f1180-0070-43bd-9ee3-095e8e2379f0/speaker/0.log" Oct 10 10:26:40 crc kubenswrapper[4669]: I1010 10:26:40.795163 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:26:40 crc kubenswrapper[4669]: E1010 10:26:40.795927 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:26:44 crc kubenswrapper[4669]: I1010 10:26:44.820734 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/util/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.075049 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/util/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.104051 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/pull/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.140461 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/pull/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.316545 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/pull/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.369182 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/util/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.381545 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/extract/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.539621 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-utilities/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.791433 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-utilities/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.791907 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-content/0.log" Oct 10 10:26:45 crc kubenswrapper[4669]: I1010 10:26:45.804669 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-content/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.037758 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-content/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.104577 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-utilities/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.361508 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-utilities/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.536682 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/registry-server/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.680761 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-content/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.693887 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-utilities/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.694415 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-content/0.log" Oct 10 10:26:46 crc kubenswrapper[4669]: I1010 10:26:46.967643 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-utilities/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.073300 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-content/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.356142 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/util/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.608422 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/pull/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.668385 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/util/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.763866 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/registry-server/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.815704 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/pull/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.946502 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/pull/0.log" Oct 10 10:26:47 crc kubenswrapper[4669]: I1010 10:26:47.982475 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/util/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.031964 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/extract/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.221324 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wf49x_90231823-44be-4740-bef0-c98a048f9b05/marketplace-operator/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.300712 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-utilities/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.543090 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-utilities/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.556813 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-content/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.576993 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-content/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.794617 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-content/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.806893 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-utilities/0.log" Oct 10 10:26:48 crc kubenswrapper[4669]: I1010 10:26:48.942325 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/registry-server/0.log" Oct 10 10:26:49 crc kubenswrapper[4669]: I1010 10:26:49.047388 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-utilities/0.log" Oct 10 10:26:49 crc kubenswrapper[4669]: I1010 10:26:49.396028 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-utilities/0.log" Oct 10 10:26:49 crc kubenswrapper[4669]: I1010 10:26:49.418190 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-content/0.log" Oct 10 10:26:49 crc kubenswrapper[4669]: I1010 10:26:49.434099 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-content/0.log" Oct 10 10:26:49 crc kubenswrapper[4669]: I1010 10:26:49.972789 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-utilities/0.log" Oct 10 10:26:49 crc kubenswrapper[4669]: I1010 10:26:49.973050 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-content/0.log" Oct 10 10:26:50 crc kubenswrapper[4669]: I1010 10:26:50.460692 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/registry-server/0.log" Oct 10 10:26:52 crc kubenswrapper[4669]: I1010 10:26:52.796095 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:26:52 crc kubenswrapper[4669]: E1010 10:26:52.796682 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:27:06 crc kubenswrapper[4669]: I1010 10:27:06.796348 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:27:06 crc kubenswrapper[4669]: E1010 10:27:06.797532 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:27:18 crc kubenswrapper[4669]: I1010 10:27:18.796668 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:27:18 crc kubenswrapper[4669]: E1010 10:27:18.797283 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:27:33 crc kubenswrapper[4669]: I1010 10:27:33.798551 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:27:34 crc kubenswrapper[4669]: I1010 10:27:34.193463 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"bc11853ac15a382d5d96f25f2a58173703d5b2ed2201893ca7b2e71e54ea802f"} Oct 10 10:29:25 crc kubenswrapper[4669]: I1010 10:29:25.155698 4669 scope.go:117] "RemoveContainer" containerID="4997f388c59ee08bd25bd04f8a4c41387712ca04cf7a544e85136c9bfa7e42b7" Oct 10 10:29:34 crc kubenswrapper[4669]: I1010 10:29:34.536465 4669 generic.go:334] "Generic (PLEG): container finished" podID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerID="82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0" exitCode=0 Oct 10 10:29:34 crc kubenswrapper[4669]: I1010 10:29:34.536542 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kpmlf/must-gather-55m2k" event={"ID":"046356c0-e374-4a1b-99bd-ff92eb5ba8e9","Type":"ContainerDied","Data":"82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0"} Oct 10 10:29:34 crc kubenswrapper[4669]: I1010 10:29:34.537911 4669 scope.go:117] "RemoveContainer" containerID="82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.345828 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kpmlf_must-gather-55m2k_046356c0-e374-4a1b-99bd-ff92eb5ba8e9/gather/0.log" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.443633 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wl8kb"] Oct 10 10:29:35 crc kubenswrapper[4669]: E1010 10:29:35.444464 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="extract-content" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.444486 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="extract-content" Oct 10 10:29:35 crc kubenswrapper[4669]: E1010 10:29:35.444504 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="extract-utilities" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.444513 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="extract-utilities" Oct 10 10:29:35 crc kubenswrapper[4669]: E1010 10:29:35.444542 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="registry-server" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.444552 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="registry-server" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.444820 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="47678fab-de95-411e-b074-ec581bc8abe1" containerName="registry-server" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.446517 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.474854 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl8kb"] Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.564902 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtpdp\" (UniqueName: \"kubernetes.io/projected/cfc4ae54-efc2-4125-a00d-78ea9f377789-kube-api-access-dtpdp\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.564964 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-catalog-content\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.565022 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-utilities\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.638217 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-56r2x"] Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.642489 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.672044 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-56r2x"] Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.687677 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-catalog-content\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.687772 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-utilities\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.687928 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtpdp\" (UniqueName: \"kubernetes.io/projected/cfc4ae54-efc2-4125-a00d-78ea9f377789-kube-api-access-dtpdp\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.688711 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-catalog-content\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.688949 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-utilities\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.733746 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtpdp\" (UniqueName: \"kubernetes.io/projected/cfc4ae54-efc2-4125-a00d-78ea9f377789-kube-api-access-dtpdp\") pod \"redhat-marketplace-wl8kb\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.790388 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-catalog-content\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.790508 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wkz5\" (UniqueName: \"kubernetes.io/projected/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-kube-api-access-9wkz5\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.790613 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-utilities\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.823740 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.916034 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wkz5\" (UniqueName: \"kubernetes.io/projected/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-kube-api-access-9wkz5\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.916277 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-utilities\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.916743 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-catalog-content\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.917310 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-catalog-content\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.917671 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-utilities\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.935275 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wkz5\" (UniqueName: \"kubernetes.io/projected/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-kube-api-access-9wkz5\") pod \"community-operators-56r2x\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:35 crc kubenswrapper[4669]: I1010 10:29:35.973275 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:36 crc kubenswrapper[4669]: I1010 10:29:36.735286 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-56r2x"] Oct 10 10:29:36 crc kubenswrapper[4669]: I1010 10:29:36.863022 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl8kb"] Oct 10 10:29:36 crc kubenswrapper[4669]: W1010 10:29:36.891956 4669 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfc4ae54_efc2_4125_a00d_78ea9f377789.slice/crio-67c654fd34cc37fb02ea194490f0b2d740913487c56943502137fe254b0350fa WatchSource:0}: Error finding container 67c654fd34cc37fb02ea194490f0b2d740913487c56943502137fe254b0350fa: Status 404 returned error can't find the container with id 67c654fd34cc37fb02ea194490f0b2d740913487c56943502137fe254b0350fa Oct 10 10:29:37 crc kubenswrapper[4669]: I1010 10:29:37.584950 4669 generic.go:334] "Generic (PLEG): container finished" podID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerID="5cc7049bf1dbd2ecb6423c1c3f443c54794097dac39e356d660e52f7be9ee5dd" exitCode=0 Oct 10 10:29:37 crc kubenswrapper[4669]: I1010 10:29:37.585066 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl8kb" event={"ID":"cfc4ae54-efc2-4125-a00d-78ea9f377789","Type":"ContainerDied","Data":"5cc7049bf1dbd2ecb6423c1c3f443c54794097dac39e356d660e52f7be9ee5dd"} Oct 10 10:29:37 crc kubenswrapper[4669]: I1010 10:29:37.585841 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl8kb" event={"ID":"cfc4ae54-efc2-4125-a00d-78ea9f377789","Type":"ContainerStarted","Data":"67c654fd34cc37fb02ea194490f0b2d740913487c56943502137fe254b0350fa"} Oct 10 10:29:37 crc kubenswrapper[4669]: I1010 10:29:37.588307 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 10:29:37 crc kubenswrapper[4669]: I1010 10:29:37.588920 4669 generic.go:334] "Generic (PLEG): container finished" podID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerID="e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be" exitCode=0 Oct 10 10:29:37 crc kubenswrapper[4669]: I1010 10:29:37.588970 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56r2x" event={"ID":"b52ac40d-5265-4d05-af0e-cb5a7f04aa91","Type":"ContainerDied","Data":"e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be"} Oct 10 10:29:37 crc kubenswrapper[4669]: I1010 10:29:37.589004 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56r2x" event={"ID":"b52ac40d-5265-4d05-af0e-cb5a7f04aa91","Type":"ContainerStarted","Data":"475cbc472587a3e5feb0e96068b6aeac8d23ee321ab02408e6624ed708641c35"} Oct 10 10:29:38 crc kubenswrapper[4669]: I1010 10:29:38.604214 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56r2x" event={"ID":"b52ac40d-5265-4d05-af0e-cb5a7f04aa91","Type":"ContainerStarted","Data":"91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e"} Oct 10 10:29:39 crc kubenswrapper[4669]: I1010 10:29:39.620046 4669 generic.go:334] "Generic (PLEG): container finished" podID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerID="c8a649ca94ee805fc05351ad7a5fe5003a8d83b2151fa0261ac6b8694cdda015" exitCode=0 Oct 10 10:29:39 crc kubenswrapper[4669]: I1010 10:29:39.620178 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl8kb" event={"ID":"cfc4ae54-efc2-4125-a00d-78ea9f377789","Type":"ContainerDied","Data":"c8a649ca94ee805fc05351ad7a5fe5003a8d83b2151fa0261ac6b8694cdda015"} Oct 10 10:29:40 crc kubenswrapper[4669]: I1010 10:29:40.633615 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl8kb" event={"ID":"cfc4ae54-efc2-4125-a00d-78ea9f377789","Type":"ContainerStarted","Data":"592368fc1f5a54f433cb5676dd072d78606b6ebb1190b8d6c553fb37b0bfbfdd"} Oct 10 10:29:40 crc kubenswrapper[4669]: I1010 10:29:40.636708 4669 generic.go:334] "Generic (PLEG): container finished" podID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerID="91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e" exitCode=0 Oct 10 10:29:40 crc kubenswrapper[4669]: I1010 10:29:40.636769 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56r2x" event={"ID":"b52ac40d-5265-4d05-af0e-cb5a7f04aa91","Type":"ContainerDied","Data":"91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e"} Oct 10 10:29:40 crc kubenswrapper[4669]: I1010 10:29:40.666179 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wl8kb" podStartSLOduration=2.897862039 podStartE2EDuration="5.666155238s" podCreationTimestamp="2025-10-10 10:29:35 +0000 UTC" firstStartedPulling="2025-10-10 10:29:37.588002968 +0000 UTC m=+4720.604021710" lastFinishedPulling="2025-10-10 10:29:40.356296167 +0000 UTC m=+4723.372314909" observedRunningTime="2025-10-10 10:29:40.664089472 +0000 UTC m=+4723.680108224" watchObservedRunningTime="2025-10-10 10:29:40.666155238 +0000 UTC m=+4723.682173980" Oct 10 10:29:41 crc kubenswrapper[4669]: I1010 10:29:41.648911 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56r2x" event={"ID":"b52ac40d-5265-4d05-af0e-cb5a7f04aa91","Type":"ContainerStarted","Data":"859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9"} Oct 10 10:29:41 crc kubenswrapper[4669]: I1010 10:29:41.681007 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-56r2x" podStartSLOduration=3.124014331 podStartE2EDuration="6.680979962s" podCreationTimestamp="2025-10-10 10:29:35 +0000 UTC" firstStartedPulling="2025-10-10 10:29:37.592892514 +0000 UTC m=+4720.608911266" lastFinishedPulling="2025-10-10 10:29:41.149858145 +0000 UTC m=+4724.165876897" observedRunningTime="2025-10-10 10:29:41.670388752 +0000 UTC m=+4724.686407494" watchObservedRunningTime="2025-10-10 10:29:41.680979962 +0000 UTC m=+4724.696998734" Oct 10 10:29:44 crc kubenswrapper[4669]: I1010 10:29:44.909907 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kpmlf/must-gather-55m2k"] Oct 10 10:29:44 crc kubenswrapper[4669]: I1010 10:29:44.911959 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-kpmlf/must-gather-55m2k" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerName="copy" containerID="cri-o://5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5" gracePeriod=2 Oct 10 10:29:44 crc kubenswrapper[4669]: I1010 10:29:44.922636 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kpmlf/must-gather-55m2k"] Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.508073 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kpmlf_must-gather-55m2k_046356c0-e374-4a1b-99bd-ff92eb5ba8e9/copy/0.log" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.508938 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.575750 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-must-gather-output\") pod \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.575891 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llff8\" (UniqueName: \"kubernetes.io/projected/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-kube-api-access-llff8\") pod \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\" (UID: \"046356c0-e374-4a1b-99bd-ff92eb5ba8e9\") " Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.587608 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-kube-api-access-llff8" (OuterVolumeSpecName: "kube-api-access-llff8") pod "046356c0-e374-4a1b-99bd-ff92eb5ba8e9" (UID: "046356c0-e374-4a1b-99bd-ff92eb5ba8e9"). InnerVolumeSpecName "kube-api-access-llff8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.678733 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llff8\" (UniqueName: \"kubernetes.io/projected/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-kube-api-access-llff8\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.693910 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kpmlf_must-gather-55m2k_046356c0-e374-4a1b-99bd-ff92eb5ba8e9/copy/0.log" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.695164 4669 generic.go:334] "Generic (PLEG): container finished" podID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerID="5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5" exitCode=143 Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.695277 4669 scope.go:117] "RemoveContainer" containerID="5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.695477 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kpmlf/must-gather-55m2k" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.723729 4669 scope.go:117] "RemoveContainer" containerID="82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.770240 4669 scope.go:117] "RemoveContainer" containerID="5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5" Oct 10 10:29:45 crc kubenswrapper[4669]: E1010 10:29:45.770937 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5\": container with ID starting with 5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5 not found: ID does not exist" containerID="5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.770986 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5"} err="failed to get container status \"5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5\": rpc error: code = NotFound desc = could not find container \"5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5\": container with ID starting with 5968e83b82459a81128f4e4eae6898336ca146ecc49a9c80aef3456df6bdb9b5 not found: ID does not exist" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.771029 4669 scope.go:117] "RemoveContainer" containerID="82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.771123 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "046356c0-e374-4a1b-99bd-ff92eb5ba8e9" (UID: "046356c0-e374-4a1b-99bd-ff92eb5ba8e9"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:29:45 crc kubenswrapper[4669]: E1010 10:29:45.771411 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0\": container with ID starting with 82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0 not found: ID does not exist" containerID="82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.771436 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0"} err="failed to get container status \"82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0\": rpc error: code = NotFound desc = could not find container \"82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0\": container with ID starting with 82ab767c723a0b94de934aad8b61cdf231bbc7830574968d02d05b92893632e0 not found: ID does not exist" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.780937 4669 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/046356c0-e374-4a1b-99bd-ff92eb5ba8e9-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.809000 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" path="/var/lib/kubelet/pods/046356c0-e374-4a1b-99bd-ff92eb5ba8e9/volumes" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.825858 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.825916 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.931529 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.976199 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:45 crc kubenswrapper[4669]: I1010 10:29:45.976262 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:46 crc kubenswrapper[4669]: I1010 10:29:46.023356 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:46 crc kubenswrapper[4669]: I1010 10:29:46.760640 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:46 crc kubenswrapper[4669]: I1010 10:29:46.761053 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:47 crc kubenswrapper[4669]: I1010 10:29:47.629836 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-56r2x"] Oct 10 10:29:48 crc kubenswrapper[4669]: I1010 10:29:48.731701 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-56r2x" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="registry-server" containerID="cri-o://859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9" gracePeriod=2 Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.034350 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl8kb"] Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.035244 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wl8kb" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="registry-server" containerID="cri-o://592368fc1f5a54f433cb5676dd072d78606b6ebb1190b8d6c553fb37b0bfbfdd" gracePeriod=2 Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.252855 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.365804 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-catalog-content\") pod \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.366846 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wkz5\" (UniqueName: \"kubernetes.io/projected/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-kube-api-access-9wkz5\") pod \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.367013 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-utilities\") pod \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\" (UID: \"b52ac40d-5265-4d05-af0e-cb5a7f04aa91\") " Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.367669 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-utilities" (OuterVolumeSpecName: "utilities") pod "b52ac40d-5265-4d05-af0e-cb5a7f04aa91" (UID: "b52ac40d-5265-4d05-af0e-cb5a7f04aa91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.378487 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-kube-api-access-9wkz5" (OuterVolumeSpecName: "kube-api-access-9wkz5") pod "b52ac40d-5265-4d05-af0e-cb5a7f04aa91" (UID: "b52ac40d-5265-4d05-af0e-cb5a7f04aa91"). InnerVolumeSpecName "kube-api-access-9wkz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.423264 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b52ac40d-5265-4d05-af0e-cb5a7f04aa91" (UID: "b52ac40d-5265-4d05-af0e-cb5a7f04aa91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.469193 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wkz5\" (UniqueName: \"kubernetes.io/projected/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-kube-api-access-9wkz5\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.469236 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.469245 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b52ac40d-5265-4d05-af0e-cb5a7f04aa91-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.751610 4669 generic.go:334] "Generic (PLEG): container finished" podID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerID="592368fc1f5a54f433cb5676dd072d78606b6ebb1190b8d6c553fb37b0bfbfdd" exitCode=0 Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.751691 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl8kb" event={"ID":"cfc4ae54-efc2-4125-a00d-78ea9f377789","Type":"ContainerDied","Data":"592368fc1f5a54f433cb5676dd072d78606b6ebb1190b8d6c553fb37b0bfbfdd"} Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.754834 4669 generic.go:334] "Generic (PLEG): container finished" podID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerID="859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9" exitCode=0 Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.754890 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56r2x" event={"ID":"b52ac40d-5265-4d05-af0e-cb5a7f04aa91","Type":"ContainerDied","Data":"859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9"} Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.754910 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-56r2x" event={"ID":"b52ac40d-5265-4d05-af0e-cb5a7f04aa91","Type":"ContainerDied","Data":"475cbc472587a3e5feb0e96068b6aeac8d23ee321ab02408e6624ed708641c35"} Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.754957 4669 scope.go:117] "RemoveContainer" containerID="859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.755209 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-56r2x" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.794922 4669 scope.go:117] "RemoveContainer" containerID="91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.833381 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-56r2x"] Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.833415 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-56r2x"] Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.847068 4669 scope.go:117] "RemoveContainer" containerID="e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.915294 4669 scope.go:117] "RemoveContainer" containerID="859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9" Oct 10 10:29:49 crc kubenswrapper[4669]: E1010 10:29:49.916283 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9\": container with ID starting with 859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9 not found: ID does not exist" containerID="859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.916326 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9"} err="failed to get container status \"859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9\": rpc error: code = NotFound desc = could not find container \"859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9\": container with ID starting with 859ec3677068a07e2a5a9e03157a4953379aca553028043096ca76dc914ed5d9 not found: ID does not exist" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.916350 4669 scope.go:117] "RemoveContainer" containerID="91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e" Oct 10 10:29:49 crc kubenswrapper[4669]: E1010 10:29:49.916948 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e\": container with ID starting with 91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e not found: ID does not exist" containerID="91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.917011 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e"} err="failed to get container status \"91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e\": rpc error: code = NotFound desc = could not find container \"91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e\": container with ID starting with 91100aa009efd5ef2783d4886415a53f7b692e8f97fcdd2959a0b5ee1fb8105e not found: ID does not exist" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.917052 4669 scope.go:117] "RemoveContainer" containerID="e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be" Oct 10 10:29:49 crc kubenswrapper[4669]: E1010 10:29:49.917442 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be\": container with ID starting with e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be not found: ID does not exist" containerID="e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be" Oct 10 10:29:49 crc kubenswrapper[4669]: I1010 10:29:49.917472 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be"} err="failed to get container status \"e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be\": rpc error: code = NotFound desc = could not find container \"e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be\": container with ID starting with e547a712f57dadbc8b4b2528890478f988509591c8626c673ae3184c754ec5be not found: ID does not exist" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.016927 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.192512 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-catalog-content\") pod \"cfc4ae54-efc2-4125-a00d-78ea9f377789\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.192686 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-utilities\") pod \"cfc4ae54-efc2-4125-a00d-78ea9f377789\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.192764 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtpdp\" (UniqueName: \"kubernetes.io/projected/cfc4ae54-efc2-4125-a00d-78ea9f377789-kube-api-access-dtpdp\") pod \"cfc4ae54-efc2-4125-a00d-78ea9f377789\" (UID: \"cfc4ae54-efc2-4125-a00d-78ea9f377789\") " Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.196515 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-utilities" (OuterVolumeSpecName: "utilities") pod "cfc4ae54-efc2-4125-a00d-78ea9f377789" (UID: "cfc4ae54-efc2-4125-a00d-78ea9f377789"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.202812 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc4ae54-efc2-4125-a00d-78ea9f377789-kube-api-access-dtpdp" (OuterVolumeSpecName: "kube-api-access-dtpdp") pod "cfc4ae54-efc2-4125-a00d-78ea9f377789" (UID: "cfc4ae54-efc2-4125-a00d-78ea9f377789"). InnerVolumeSpecName "kube-api-access-dtpdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.209000 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfc4ae54-efc2-4125-a00d-78ea9f377789" (UID: "cfc4ae54-efc2-4125-a00d-78ea9f377789"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.295321 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.295353 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfc4ae54-efc2-4125-a00d-78ea9f377789-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.295421 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtpdp\" (UniqueName: \"kubernetes.io/projected/cfc4ae54-efc2-4125-a00d-78ea9f377789-kube-api-access-dtpdp\") on node \"crc\" DevicePath \"\"" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.771957 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wl8kb" event={"ID":"cfc4ae54-efc2-4125-a00d-78ea9f377789","Type":"ContainerDied","Data":"67c654fd34cc37fb02ea194490f0b2d740913487c56943502137fe254b0350fa"} Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.772372 4669 scope.go:117] "RemoveContainer" containerID="592368fc1f5a54f433cb5676dd072d78606b6ebb1190b8d6c553fb37b0bfbfdd" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.772008 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wl8kb" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.803847 4669 scope.go:117] "RemoveContainer" containerID="c8a649ca94ee805fc05351ad7a5fe5003a8d83b2151fa0261ac6b8694cdda015" Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.833733 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl8kb"] Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.846902 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wl8kb"] Oct 10 10:29:50 crc kubenswrapper[4669]: I1010 10:29:50.849046 4669 scope.go:117] "RemoveContainer" containerID="5cc7049bf1dbd2ecb6423c1c3f443c54794097dac39e356d660e52f7be9ee5dd" Oct 10 10:29:51 crc kubenswrapper[4669]: I1010 10:29:51.806572 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" path="/var/lib/kubelet/pods/b52ac40d-5265-4d05-af0e-cb5a7f04aa91/volumes" Oct 10 10:29:51 crc kubenswrapper[4669]: I1010 10:29:51.808228 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" path="/var/lib/kubelet/pods/cfc4ae54-efc2-4125-a00d-78ea9f377789/volumes" Oct 10 10:29:54 crc kubenswrapper[4669]: I1010 10:29:54.275419 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:29:54 crc kubenswrapper[4669]: I1010 10:29:54.276701 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.174131 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q"] Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.176428 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerName="gather" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.176895 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerName="gather" Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.176979 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerName="copy" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.177057 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerName="copy" Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.177138 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="extract-utilities" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.177209 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="extract-utilities" Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.177318 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="registry-server" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.177389 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="registry-server" Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.177468 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="registry-server" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.177537 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="registry-server" Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.177642 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="extract-utilities" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.177866 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="extract-utilities" Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.177959 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="extract-content" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.178039 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="extract-content" Oct 10 10:30:00 crc kubenswrapper[4669]: E1010 10:30:00.178134 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="extract-content" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.178204 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="extract-content" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.178520 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerName="gather" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.178644 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="b52ac40d-5265-4d05-af0e-cb5a7f04aa91" containerName="registry-server" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.178781 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc4ae54-efc2-4125-a00d-78ea9f377789" containerName="registry-server" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.178886 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="046356c0-e374-4a1b-99bd-ff92eb5ba8e9" containerName="copy" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.179848 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.182527 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.182718 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.188160 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q"] Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.324701 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-config-volume\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.325052 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-secret-volume\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.325162 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpc5v\" (UniqueName: \"kubernetes.io/projected/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-kube-api-access-lpc5v\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.426661 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpc5v\" (UniqueName: \"kubernetes.io/projected/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-kube-api-access-lpc5v\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.426817 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-config-volume\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.426897 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-secret-volume\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.428118 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-config-volume\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.434578 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-secret-volume\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.444814 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpc5v\" (UniqueName: \"kubernetes.io/projected/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-kube-api-access-lpc5v\") pod \"collect-profiles-29334870-rc96q\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:00 crc kubenswrapper[4669]: I1010 10:30:00.517535 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:01 crc kubenswrapper[4669]: I1010 10:30:01.082578 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q"] Oct 10 10:30:01 crc kubenswrapper[4669]: I1010 10:30:01.887546 4669 generic.go:334] "Generic (PLEG): container finished" podID="6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3" containerID="5a41d561d4f4ea2348ea2b7a367fb96e8ecb29cda69dbc5c8b255e81b4c9b48e" exitCode=0 Oct 10 10:30:01 crc kubenswrapper[4669]: I1010 10:30:01.887663 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" event={"ID":"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3","Type":"ContainerDied","Data":"5a41d561d4f4ea2348ea2b7a367fb96e8ecb29cda69dbc5c8b255e81b4c9b48e"} Oct 10 10:30:01 crc kubenswrapper[4669]: I1010 10:30:01.888050 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" event={"ID":"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3","Type":"ContainerStarted","Data":"0743564522d75e9a7cfa3b7b9934dfd1f8087cd273b2eb5f9090a035c71912d5"} Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.448090 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.598961 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-secret-volume\") pod \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.599082 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-config-volume\") pod \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.599269 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpc5v\" (UniqueName: \"kubernetes.io/projected/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-kube-api-access-lpc5v\") pod \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\" (UID: \"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3\") " Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.600184 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-config-volume" (OuterVolumeSpecName: "config-volume") pod "6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3" (UID: "6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.606193 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3" (UID: "6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.607134 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-kube-api-access-lpc5v" (OuterVolumeSpecName: "kube-api-access-lpc5v") pod "6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3" (UID: "6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3"). InnerVolumeSpecName "kube-api-access-lpc5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.702253 4669 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.702304 4669 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.702316 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpc5v\" (UniqueName: \"kubernetes.io/projected/6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3-kube-api-access-lpc5v\") on node \"crc\" DevicePath \"\"" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.913691 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" event={"ID":"6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3","Type":"ContainerDied","Data":"0743564522d75e9a7cfa3b7b9934dfd1f8087cd273b2eb5f9090a035c71912d5"} Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.914123 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0743564522d75e9a7cfa3b7b9934dfd1f8087cd273b2eb5f9090a035c71912d5" Oct 10 10:30:03 crc kubenswrapper[4669]: I1010 10:30:03.913749 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29334870-rc96q" Oct 10 10:30:04 crc kubenswrapper[4669]: I1010 10:30:04.541521 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46"] Oct 10 10:30:04 crc kubenswrapper[4669]: I1010 10:30:04.549667 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29334825-7sv46"] Oct 10 10:30:05 crc kubenswrapper[4669]: I1010 10:30:05.814329 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1442cabb-eb6c-4a8a-9c00-d272bdecc2f0" path="/var/lib/kubelet/pods/1442cabb-eb6c-4a8a-9c00-d272bdecc2f0/volumes" Oct 10 10:30:24 crc kubenswrapper[4669]: I1010 10:30:24.275105 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:30:24 crc kubenswrapper[4669]: I1010 10:30:24.275985 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.273097 4669 scope.go:117] "RemoveContainer" containerID="75503a8cddbf3b864ad271e340a446e3e4ae36027424a41ae29a4513158b7116" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.837451 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xkr7t/must-gather-pzqxv"] Oct 10 10:30:25 crc kubenswrapper[4669]: E1010 10:30:25.839749 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3" containerName="collect-profiles" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.839846 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3" containerName="collect-profiles" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.840121 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef21d91-bc1b-4fa8-935f-c2fc9af1acb3" containerName="collect-profiles" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.841824 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.850396 4669 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xkr7t"/"default-dockercfg-8dkkm" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.851202 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xkr7t"/"kube-root-ca.crt" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.851341 4669 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xkr7t"/"openshift-service-ca.crt" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.865294 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xkr7t/must-gather-pzqxv"] Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.949567 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d49fc2f9-06e0-409c-9829-65375fad7e5b-must-gather-output\") pod \"must-gather-pzqxv\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:25 crc kubenswrapper[4669]: I1010 10:30:25.949719 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj9wx\" (UniqueName: \"kubernetes.io/projected/d49fc2f9-06e0-409c-9829-65375fad7e5b-kube-api-access-zj9wx\") pod \"must-gather-pzqxv\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:26 crc kubenswrapper[4669]: I1010 10:30:26.052291 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj9wx\" (UniqueName: \"kubernetes.io/projected/d49fc2f9-06e0-409c-9829-65375fad7e5b-kube-api-access-zj9wx\") pod \"must-gather-pzqxv\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:26 crc kubenswrapper[4669]: I1010 10:30:26.053020 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d49fc2f9-06e0-409c-9829-65375fad7e5b-must-gather-output\") pod \"must-gather-pzqxv\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:26 crc kubenswrapper[4669]: I1010 10:30:26.053721 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d49fc2f9-06e0-409c-9829-65375fad7e5b-must-gather-output\") pod \"must-gather-pzqxv\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:26 crc kubenswrapper[4669]: I1010 10:30:26.073749 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj9wx\" (UniqueName: \"kubernetes.io/projected/d49fc2f9-06e0-409c-9829-65375fad7e5b-kube-api-access-zj9wx\") pod \"must-gather-pzqxv\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:26 crc kubenswrapper[4669]: I1010 10:30:26.202523 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:30:26 crc kubenswrapper[4669]: I1010 10:30:26.741728 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xkr7t/must-gather-pzqxv"] Oct 10 10:30:27 crc kubenswrapper[4669]: I1010 10:30:27.174746 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" event={"ID":"d49fc2f9-06e0-409c-9829-65375fad7e5b","Type":"ContainerStarted","Data":"e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0"} Oct 10 10:30:27 crc kubenswrapper[4669]: I1010 10:30:27.175185 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" event={"ID":"d49fc2f9-06e0-409c-9829-65375fad7e5b","Type":"ContainerStarted","Data":"9058563d688c5ff3a36d3bfae29495a097724cd2698d8752535c4ae891a1fac4"} Oct 10 10:30:28 crc kubenswrapper[4669]: I1010 10:30:28.186018 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" event={"ID":"d49fc2f9-06e0-409c-9829-65375fad7e5b","Type":"ContainerStarted","Data":"0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34"} Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.545001 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" podStartSLOduration=8.54497741 podStartE2EDuration="8.54497741s" podCreationTimestamp="2025-10-10 10:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:30:28.204056546 +0000 UTC m=+4771.220075288" watchObservedRunningTime="2025-10-10 10:30:33.54497741 +0000 UTC m=+4776.560996152" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.553596 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-d46bs"] Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.555118 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.633724 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgglz\" (UniqueName: \"kubernetes.io/projected/a3a38678-3245-40d7-b63f-d04c83e301e9-kube-api-access-rgglz\") pod \"crc-debug-d46bs\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.633950 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a3a38678-3245-40d7-b63f-d04c83e301e9-host\") pod \"crc-debug-d46bs\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.735571 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgglz\" (UniqueName: \"kubernetes.io/projected/a3a38678-3245-40d7-b63f-d04c83e301e9-kube-api-access-rgglz\") pod \"crc-debug-d46bs\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.735887 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a3a38678-3245-40d7-b63f-d04c83e301e9-host\") pod \"crc-debug-d46bs\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.736754 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a3a38678-3245-40d7-b63f-d04c83e301e9-host\") pod \"crc-debug-d46bs\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.760020 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgglz\" (UniqueName: \"kubernetes.io/projected/a3a38678-3245-40d7-b63f-d04c83e301e9-kube-api-access-rgglz\") pod \"crc-debug-d46bs\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:33 crc kubenswrapper[4669]: I1010 10:30:33.883030 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:30:34 crc kubenswrapper[4669]: I1010 10:30:34.257961 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" event={"ID":"a3a38678-3245-40d7-b63f-d04c83e301e9","Type":"ContainerStarted","Data":"7d726e68c009b93af97c9effbba36df188fe8b75fc3b1835d1a0af37491b0246"} Oct 10 10:30:34 crc kubenswrapper[4669]: I1010 10:30:34.258401 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" event={"ID":"a3a38678-3245-40d7-b63f-d04c83e301e9","Type":"ContainerStarted","Data":"db4b8768f64f5817504ffd9c984ec94c618c759c0d5d78ded1d63629500e1b20"} Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.274825 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.275645 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.275700 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.276630 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc11853ac15a382d5d96f25f2a58173703d5b2ed2201893ca7b2e71e54ea802f"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.276710 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://bc11853ac15a382d5d96f25f2a58173703d5b2ed2201893ca7b2e71e54ea802f" gracePeriod=600 Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.491140 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="bc11853ac15a382d5d96f25f2a58173703d5b2ed2201893ca7b2e71e54ea802f" exitCode=0 Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.491212 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"bc11853ac15a382d5d96f25f2a58173703d5b2ed2201893ca7b2e71e54ea802f"} Oct 10 10:30:54 crc kubenswrapper[4669]: I1010 10:30:54.491526 4669 scope.go:117] "RemoveContainer" containerID="91ee10a9f989f35c963caea9a0479607f97d660ae36692055fc620c7d8fd3d29" Oct 10 10:30:56 crc kubenswrapper[4669]: I1010 10:30:56.530336 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d"} Oct 10 10:30:56 crc kubenswrapper[4669]: I1010 10:30:56.555026 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" podStartSLOduration=23.55499484 podStartE2EDuration="23.55499484s" podCreationTimestamp="2025-10-10 10:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:30:34.282517105 +0000 UTC m=+4777.298535847" watchObservedRunningTime="2025-10-10 10:30:56.55499484 +0000 UTC m=+4799.571013582" Oct 10 10:31:19 crc kubenswrapper[4669]: I1010 10:31:19.757457 4669 generic.go:334] "Generic (PLEG): container finished" podID="a3a38678-3245-40d7-b63f-d04c83e301e9" containerID="7d726e68c009b93af97c9effbba36df188fe8b75fc3b1835d1a0af37491b0246" exitCode=0 Oct 10 10:31:19 crc kubenswrapper[4669]: I1010 10:31:19.757565 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" event={"ID":"a3a38678-3245-40d7-b63f-d04c83e301e9","Type":"ContainerDied","Data":"7d726e68c009b93af97c9effbba36df188fe8b75fc3b1835d1a0af37491b0246"} Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.890034 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.930857 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-d46bs"] Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.942150 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-d46bs"] Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.968080 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a3a38678-3245-40d7-b63f-d04c83e301e9-host\") pod \"a3a38678-3245-40d7-b63f-d04c83e301e9\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.968177 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3a38678-3245-40d7-b63f-d04c83e301e9-host" (OuterVolumeSpecName: "host") pod "a3a38678-3245-40d7-b63f-d04c83e301e9" (UID: "a3a38678-3245-40d7-b63f-d04c83e301e9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.968514 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgglz\" (UniqueName: \"kubernetes.io/projected/a3a38678-3245-40d7-b63f-d04c83e301e9-kube-api-access-rgglz\") pod \"a3a38678-3245-40d7-b63f-d04c83e301e9\" (UID: \"a3a38678-3245-40d7-b63f-d04c83e301e9\") " Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.970295 4669 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a3a38678-3245-40d7-b63f-d04c83e301e9-host\") on node \"crc\" DevicePath \"\"" Oct 10 10:31:20 crc kubenswrapper[4669]: I1010 10:31:20.975403 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a38678-3245-40d7-b63f-d04c83e301e9-kube-api-access-rgglz" (OuterVolumeSpecName: "kube-api-access-rgglz") pod "a3a38678-3245-40d7-b63f-d04c83e301e9" (UID: "a3a38678-3245-40d7-b63f-d04c83e301e9"). InnerVolumeSpecName "kube-api-access-rgglz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:31:21 crc kubenswrapper[4669]: I1010 10:31:21.073042 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgglz\" (UniqueName: \"kubernetes.io/projected/a3a38678-3245-40d7-b63f-d04c83e301e9-kube-api-access-rgglz\") on node \"crc\" DevicePath \"\"" Oct 10 10:31:21 crc kubenswrapper[4669]: I1010 10:31:21.792394 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db4b8768f64f5817504ffd9c984ec94c618c759c0d5d78ded1d63629500e1b20" Oct 10 10:31:21 crc kubenswrapper[4669]: I1010 10:31:21.792486 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-d46bs" Oct 10 10:31:21 crc kubenswrapper[4669]: I1010 10:31:21.816891 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a38678-3245-40d7-b63f-d04c83e301e9" path="/var/lib/kubelet/pods/a3a38678-3245-40d7-b63f-d04c83e301e9/volumes" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.193990 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-blzpz"] Oct 10 10:31:22 crc kubenswrapper[4669]: E1010 10:31:22.199310 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a38678-3245-40d7-b63f-d04c83e301e9" containerName="container-00" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.199555 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a38678-3245-40d7-b63f-d04c83e301e9" containerName="container-00" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.199950 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a38678-3245-40d7-b63f-d04c83e301e9" containerName="container-00" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.200951 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.302062 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d858917b-7737-4c4d-96e9-0b3ea74a23cf-host\") pod \"crc-debug-blzpz\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.302105 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvffz\" (UniqueName: \"kubernetes.io/projected/d858917b-7737-4c4d-96e9-0b3ea74a23cf-kube-api-access-xvffz\") pod \"crc-debug-blzpz\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.404462 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d858917b-7737-4c4d-96e9-0b3ea74a23cf-host\") pod \"crc-debug-blzpz\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.404516 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvffz\" (UniqueName: \"kubernetes.io/projected/d858917b-7737-4c4d-96e9-0b3ea74a23cf-kube-api-access-xvffz\") pod \"crc-debug-blzpz\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.404629 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d858917b-7737-4c4d-96e9-0b3ea74a23cf-host\") pod \"crc-debug-blzpz\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.425663 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvffz\" (UniqueName: \"kubernetes.io/projected/d858917b-7737-4c4d-96e9-0b3ea74a23cf-kube-api-access-xvffz\") pod \"crc-debug-blzpz\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.524976 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.809543 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" event={"ID":"d858917b-7737-4c4d-96e9-0b3ea74a23cf","Type":"ContainerStarted","Data":"eb807a1160c7991e15ecfdba14eb03746743a184a324702d5af4b70914b89b46"} Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.809924 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" event={"ID":"d858917b-7737-4c4d-96e9-0b3ea74a23cf","Type":"ContainerStarted","Data":"01694593ed21ca52e226e83875c498778267aba503edb06230e201ba29ce24f6"} Oct 10 10:31:22 crc kubenswrapper[4669]: I1010 10:31:22.890518 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" podStartSLOduration=0.890491931 podStartE2EDuration="890.491931ms" podCreationTimestamp="2025-10-10 10:31:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:31:22.834782277 +0000 UTC m=+4825.850801019" watchObservedRunningTime="2025-10-10 10:31:22.890491931 +0000 UTC m=+4825.906510673" Oct 10 10:31:23 crc kubenswrapper[4669]: I1010 10:31:23.820973 4669 generic.go:334] "Generic (PLEG): container finished" podID="d858917b-7737-4c4d-96e9-0b3ea74a23cf" containerID="eb807a1160c7991e15ecfdba14eb03746743a184a324702d5af4b70914b89b46" exitCode=0 Oct 10 10:31:23 crc kubenswrapper[4669]: I1010 10:31:23.821147 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" event={"ID":"d858917b-7737-4c4d-96e9-0b3ea74a23cf","Type":"ContainerDied","Data":"eb807a1160c7991e15ecfdba14eb03746743a184a324702d5af4b70914b89b46"} Oct 10 10:31:24 crc kubenswrapper[4669]: I1010 10:31:24.968994 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.008449 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-blzpz"] Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.016460 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-blzpz"] Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.061622 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d858917b-7737-4c4d-96e9-0b3ea74a23cf-host\") pod \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.061707 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvffz\" (UniqueName: \"kubernetes.io/projected/d858917b-7737-4c4d-96e9-0b3ea74a23cf-kube-api-access-xvffz\") pod \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\" (UID: \"d858917b-7737-4c4d-96e9-0b3ea74a23cf\") " Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.061745 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d858917b-7737-4c4d-96e9-0b3ea74a23cf-host" (OuterVolumeSpecName: "host") pod "d858917b-7737-4c4d-96e9-0b3ea74a23cf" (UID: "d858917b-7737-4c4d-96e9-0b3ea74a23cf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.062480 4669 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d858917b-7737-4c4d-96e9-0b3ea74a23cf-host\") on node \"crc\" DevicePath \"\"" Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.070409 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d858917b-7737-4c4d-96e9-0b3ea74a23cf-kube-api-access-xvffz" (OuterVolumeSpecName: "kube-api-access-xvffz") pod "d858917b-7737-4c4d-96e9-0b3ea74a23cf" (UID: "d858917b-7737-4c4d-96e9-0b3ea74a23cf"). InnerVolumeSpecName "kube-api-access-xvffz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.165156 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvffz\" (UniqueName: \"kubernetes.io/projected/d858917b-7737-4c4d-96e9-0b3ea74a23cf-kube-api-access-xvffz\") on node \"crc\" DevicePath \"\"" Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.810491 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d858917b-7737-4c4d-96e9-0b3ea74a23cf" path="/var/lib/kubelet/pods/d858917b-7737-4c4d-96e9-0b3ea74a23cf/volumes" Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.841313 4669 scope.go:117] "RemoveContainer" containerID="eb807a1160c7991e15ecfdba14eb03746743a184a324702d5af4b70914b89b46" Oct 10 10:31:25 crc kubenswrapper[4669]: I1010 10:31:25.841461 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-blzpz" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.244932 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-mc27s"] Oct 10 10:31:26 crc kubenswrapper[4669]: E1010 10:31:26.245961 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d858917b-7737-4c4d-96e9-0b3ea74a23cf" containerName="container-00" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.245980 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="d858917b-7737-4c4d-96e9-0b3ea74a23cf" containerName="container-00" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.246160 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="d858917b-7737-4c4d-96e9-0b3ea74a23cf" containerName="container-00" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.246899 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.298749 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/362ce889-52b7-42ba-8ee5-306532d627c1-host\") pod \"crc-debug-mc27s\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.298845 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqvv8\" (UniqueName: \"kubernetes.io/projected/362ce889-52b7-42ba-8ee5-306532d627c1-kube-api-access-zqvv8\") pod \"crc-debug-mc27s\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.401558 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/362ce889-52b7-42ba-8ee5-306532d627c1-host\") pod \"crc-debug-mc27s\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.401669 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqvv8\" (UniqueName: \"kubernetes.io/projected/362ce889-52b7-42ba-8ee5-306532d627c1-kube-api-access-zqvv8\") pod \"crc-debug-mc27s\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.402632 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/362ce889-52b7-42ba-8ee5-306532d627c1-host\") pod \"crc-debug-mc27s\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.431056 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqvv8\" (UniqueName: \"kubernetes.io/projected/362ce889-52b7-42ba-8ee5-306532d627c1-kube-api-access-zqvv8\") pod \"crc-debug-mc27s\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.574042 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.853927 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" event={"ID":"362ce889-52b7-42ba-8ee5-306532d627c1","Type":"ContainerStarted","Data":"038a1f2f682182bf07ae7bd14626714d8ae713b56c4fc0d2913471592afa23eb"} Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.854327 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" event={"ID":"362ce889-52b7-42ba-8ee5-306532d627c1","Type":"ContainerStarted","Data":"0b14d6544f9eccca942686cdff835b8b10fc145c2863ca8558cc09c8b9724a33"} Oct 10 10:31:26 crc kubenswrapper[4669]: I1010 10:31:26.876228 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" podStartSLOduration=0.876209881 podStartE2EDuration="876.209881ms" podCreationTimestamp="2025-10-10 10:31:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 10:31:26.870035463 +0000 UTC m=+4829.886054205" watchObservedRunningTime="2025-10-10 10:31:26.876209881 +0000 UTC m=+4829.892228623" Oct 10 10:31:27 crc kubenswrapper[4669]: I1010 10:31:27.870939 4669 generic.go:334] "Generic (PLEG): container finished" podID="362ce889-52b7-42ba-8ee5-306532d627c1" containerID="038a1f2f682182bf07ae7bd14626714d8ae713b56c4fc0d2913471592afa23eb" exitCode=0 Oct 10 10:31:27 crc kubenswrapper[4669]: I1010 10:31:27.871021 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" event={"ID":"362ce889-52b7-42ba-8ee5-306532d627c1","Type":"ContainerDied","Data":"038a1f2f682182bf07ae7bd14626714d8ae713b56c4fc0d2913471592afa23eb"} Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.018534 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.062416 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqvv8\" (UniqueName: \"kubernetes.io/projected/362ce889-52b7-42ba-8ee5-306532d627c1-kube-api-access-zqvv8\") pod \"362ce889-52b7-42ba-8ee5-306532d627c1\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.062840 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/362ce889-52b7-42ba-8ee5-306532d627c1-host\") pod \"362ce889-52b7-42ba-8ee5-306532d627c1\" (UID: \"362ce889-52b7-42ba-8ee5-306532d627c1\") " Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.063854 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/362ce889-52b7-42ba-8ee5-306532d627c1-host" (OuterVolumeSpecName: "host") pod "362ce889-52b7-42ba-8ee5-306532d627c1" (UID: "362ce889-52b7-42ba-8ee5-306532d627c1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.065756 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-mc27s"] Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.087825 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362ce889-52b7-42ba-8ee5-306532d627c1-kube-api-access-zqvv8" (OuterVolumeSpecName: "kube-api-access-zqvv8") pod "362ce889-52b7-42ba-8ee5-306532d627c1" (UID: "362ce889-52b7-42ba-8ee5-306532d627c1"). InnerVolumeSpecName "kube-api-access-zqvv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.121049 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xkr7t/crc-debug-mc27s"] Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.165799 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqvv8\" (UniqueName: \"kubernetes.io/projected/362ce889-52b7-42ba-8ee5-306532d627c1-kube-api-access-zqvv8\") on node \"crc\" DevicePath \"\"" Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.165844 4669 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/362ce889-52b7-42ba-8ee5-306532d627c1-host\") on node \"crc\" DevicePath \"\"" Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.812684 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="362ce889-52b7-42ba-8ee5-306532d627c1" path="/var/lib/kubelet/pods/362ce889-52b7-42ba-8ee5-306532d627c1/volumes" Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.894369 4669 scope.go:117] "RemoveContainer" containerID="038a1f2f682182bf07ae7bd14626714d8ae713b56c4fc0d2913471592afa23eb" Oct 10 10:31:29 crc kubenswrapper[4669]: I1010 10:31:29.894892 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/crc-debug-mc27s" Oct 10 10:32:19 crc kubenswrapper[4669]: I1010 10:32:19.336043 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8c86b488-lgxwv_9306ad2c-0c67-4a1a-8031-19d4cddce8d0/barbican-api/0.log" Oct 10 10:32:19 crc kubenswrapper[4669]: I1010 10:32:19.505538 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8c86b488-lgxwv_9306ad2c-0c67-4a1a-8031-19d4cddce8d0/barbican-api-log/0.log" Oct 10 10:32:19 crc kubenswrapper[4669]: I1010 10:32:19.640300 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fcbcfb6bb-9hc2p_bb7dc377-170c-4504-a61b-386110e80526/barbican-keystone-listener/0.log" Oct 10 10:32:19 crc kubenswrapper[4669]: I1010 10:32:19.698204 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fcbcfb6bb-9hc2p_bb7dc377-170c-4504-a61b-386110e80526/barbican-keystone-listener-log/0.log" Oct 10 10:32:19 crc kubenswrapper[4669]: I1010 10:32:19.920309 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-686db9dfc7-hg7z4_58f4d796-043f-48f0-a649-99fb297c694a/barbican-worker-log/0.log" Oct 10 10:32:19 crc kubenswrapper[4669]: I1010 10:32:19.935623 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-686db9dfc7-hg7z4_58f4d796-043f-48f0-a649-99fb297c694a/barbican-worker/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.119089 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gdw9v_66dbecf2-f169-4c6c-b8be-efbab612ace8/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.212552 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/ceilometer-central-agent/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.363947 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/ceilometer-notification-agent/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.438035 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/proxy-httpd/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.502834 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_586ee21d-4f34-48fe-8528-fe925bb8e35a/sg-core/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.598374 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-bkd4q_c1b45643-fd51-4a37-a298-f9b84b997c40/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.735234 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-rzbgv_e53d40a4-c456-4991-8cfc-4dc303b5205d/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.948100 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50ffa58f-0d75-4996-b66c-092aacc84c63/cinder-api/0.log" Oct 10 10:32:20 crc kubenswrapper[4669]: I1010 10:32:20.975724 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_50ffa58f-0d75-4996-b66c-092aacc84c63/cinder-api-log/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.232792 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c642048f-8f7f-4355-bd06-e862839e25e9/cinder-backup/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.252296 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c642048f-8f7f-4355-bd06-e862839e25e9/probe/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.480440 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23eadb50-c141-42eb-b24c-2a493eec134e/cinder-scheduler/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.518841 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23eadb50-c141-42eb-b24c-2a493eec134e/probe/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.661195 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3c92ae00-f31f-4d27-aaca-8d3920a345eb/cinder-volume/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.717874 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3c92ae00-f31f-4d27-aaca-8d3920a345eb/probe/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.929167 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-st5n8_2d5855cf-c66a-4daa-9b94-84636ee8936d/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:21 crc kubenswrapper[4669]: I1010 10:32:21.934018 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-hqbdg_7469da7f-e608-439c-93b1-d60130d75a72/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:22 crc kubenswrapper[4669]: I1010 10:32:22.183539 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58d5f4b747-fl954_2dff0743-e2de-497b-8b6e-8b7773e19da5/init/0.log" Oct 10 10:32:22 crc kubenswrapper[4669]: I1010 10:32:22.522560 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58d5f4b747-fl954_2dff0743-e2de-497b-8b6e-8b7773e19da5/init/0.log" Oct 10 10:32:22 crc kubenswrapper[4669]: I1010 10:32:22.614725 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_60e2fea4-35a0-45d6-bd36-e7172b6822fe/glance-httpd/0.log" Oct 10 10:32:22 crc kubenswrapper[4669]: I1010 10:32:22.671888 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-58d5f4b747-fl954_2dff0743-e2de-497b-8b6e-8b7773e19da5/dnsmasq-dns/0.log" Oct 10 10:32:22 crc kubenswrapper[4669]: I1010 10:32:22.910630 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_731fc6e9-0f93-4e9f-bbc0-767b6fada47b/glance-log/0.log" Oct 10 10:32:22 crc kubenswrapper[4669]: I1010 10:32:22.927806 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_731fc6e9-0f93-4e9f-bbc0-767b6fada47b/glance-httpd/0.log" Oct 10 10:32:22 crc kubenswrapper[4669]: I1010 10:32:22.970730 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_60e2fea4-35a0-45d6-bd36-e7172b6822fe/glance-log/0.log" Oct 10 10:32:23 crc kubenswrapper[4669]: I1010 10:32:23.420291 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-566f97fbdd-plcxg_0040911a-8dff-49e4-99af-aafb84f7639a/horizon/0.log" Oct 10 10:32:23 crc kubenswrapper[4669]: I1010 10:32:23.577846 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-566f97fbdd-plcxg_0040911a-8dff-49e4-99af-aafb84f7639a/horizon-log/0.log" Oct 10 10:32:23 crc kubenswrapper[4669]: I1010 10:32:23.736650 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hzlnz_9b61fc7a-573d-4a8d-bdbc-08c05cd5bb6f/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:23 crc kubenswrapper[4669]: I1010 10:32:23.776558 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-qxr8b_880204d3-a81a-433f-ace4-ec7ff0092b53/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:23 crc kubenswrapper[4669]: I1010 10:32:23.980938 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29334841-rb65h_d03bcfa8-9d28-46ac-b0aa-1e56ba6bb759/keystone-cron/0.log" Oct 10 10:32:23 crc kubenswrapper[4669]: I1010 10:32:23.991109 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-65b8b49f6-vm5h8_8c9b81a2-843b-4bab-b58a-5e2a1ce2e636/keystone-api/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.097007 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_0330c42a-ef18-46cd-9a95-5622b53328b7/kube-state-metrics/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.351733 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-8kd62_fbaa41c7-694f-4fda-aaf5-481188158475/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.476038 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_53ead5e1-28e3-4677-99e0-e2faa30ae9a5/manila-api/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.494463 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_53ead5e1-28e3-4677-99e0-e2faa30ae9a5/manila-api-log/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.701311 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_8a1791d4-98d8-4eac-9676-e409ac975611/probe/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.709768 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_8a1791d4-98d8-4eac-9676-e409ac975611/manila-scheduler/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.846618 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e4865c42-8dbc-4d96-afb9-e65d85f8ec96/manila-share/0.log" Oct 10 10:32:24 crc kubenswrapper[4669]: I1010 10:32:24.976231 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_e4865c42-8dbc-4d96-afb9-e65d85f8ec96/probe/0.log" Oct 10 10:32:25 crc kubenswrapper[4669]: I1010 10:32:25.288060 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6d556ffc9f-2x4qt_5bf9895c-806b-4a83-ae94-7530013432fb/neutron-httpd/0.log" Oct 10 10:32:25 crc kubenswrapper[4669]: I1010 10:32:25.290525 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6d556ffc9f-2x4qt_5bf9895c-806b-4a83-ae94-7530013432fb/neutron-api/0.log" Oct 10 10:32:25 crc kubenswrapper[4669]: I1010 10:32:25.591829 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fw25w_f9d4c45c-5b27-4012-a0d0-72a1a2b1843e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:26 crc kubenswrapper[4669]: I1010 10:32:26.158470 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d81e619a-c272-4824-b19f-dea88a428acd/nova-api-log/0.log" Oct 10 10:32:26 crc kubenswrapper[4669]: I1010 10:32:26.278329 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0c654342-12f4-40f3-a4dd-e4e5a76b8b76/nova-cell0-conductor-conductor/0.log" Oct 10 10:32:26 crc kubenswrapper[4669]: I1010 10:32:26.737692 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_9a6e22fc-ac5c-469b-bf5e-3ce1de48968a/nova-cell1-conductor-conductor/0.log" Oct 10 10:32:26 crc kubenswrapper[4669]: I1010 10:32:26.814362 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5d5e59af-66c7-4723-9758-9369ccf46cde/nova-cell1-novncproxy-novncproxy/0.log" Oct 10 10:32:26 crc kubenswrapper[4669]: I1010 10:32:26.827001 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d81e619a-c272-4824-b19f-dea88a428acd/nova-api-api/0.log" Oct 10 10:32:27 crc kubenswrapper[4669]: I1010 10:32:27.217821 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-ncrw6_4933ae0b-dda6-44e1-a551-730e803bb7d5/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:27 crc kubenswrapper[4669]: I1010 10:32:27.310481 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f6b1ea56-8a03-4b04-ba07-0852d3fb6545/nova-metadata-log/0.log" Oct 10 10:32:27 crc kubenswrapper[4669]: I1010 10:32:27.831817 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_02229a1f-8349-4e0c-b5b5-b56ba246f395/mysql-bootstrap/0.log" Oct 10 10:32:27 crc kubenswrapper[4669]: I1010 10:32:27.837724 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_cbfa50a5-b829-4023-b978-d76660fc77b2/nova-scheduler-scheduler/0.log" Oct 10 10:32:28 crc kubenswrapper[4669]: I1010 10:32:28.019746 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_02229a1f-8349-4e0c-b5b5-b56ba246f395/mysql-bootstrap/0.log" Oct 10 10:32:28 crc kubenswrapper[4669]: I1010 10:32:28.123918 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_02229a1f-8349-4e0c-b5b5-b56ba246f395/galera/0.log" Oct 10 10:32:28 crc kubenswrapper[4669]: I1010 10:32:28.332010 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_79d402e9-8187-4cd5-9df5-9297d8fc4b5a/mysql-bootstrap/0.log" Oct 10 10:32:28 crc kubenswrapper[4669]: I1010 10:32:28.576506 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_79d402e9-8187-4cd5-9df5-9297d8fc4b5a/galera/0.log" Oct 10 10:32:28 crc kubenswrapper[4669]: I1010 10:32:28.641285 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_79d402e9-8187-4cd5-9df5-9297d8fc4b5a/mysql-bootstrap/0.log" Oct 10 10:32:28 crc kubenswrapper[4669]: I1010 10:32:28.863852 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2bb8c92a-28c8-4424-94c2-c577f0b6b486/openstackclient/0.log" Oct 10 10:32:29 crc kubenswrapper[4669]: I1010 10:32:29.036780 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jrw5q_2b815c31-96c7-47c3-b458-35945b55b326/ovn-controller/0.log" Oct 10 10:32:29 crc kubenswrapper[4669]: I1010 10:32:29.317072 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f6b1ea56-8a03-4b04-ba07-0852d3fb6545/nova-metadata-metadata/0.log" Oct 10 10:32:29 crc kubenswrapper[4669]: I1010 10:32:29.609353 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vg62q_1f2fdb85-cfc4-471d-b35e-9388c3406895/openstack-network-exporter/0.log" Oct 10 10:32:29 crc kubenswrapper[4669]: I1010 10:32:29.828786 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovsdb-server-init/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.071236 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovs-vswitchd/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.113720 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovsdb-server/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.148484 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-d8wg9_db075709-6fea-46c4-b711-c5517eaae350/ovsdb-server-init/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.360336 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-nsc5k_cd3df673-494f-4e69-a140-34e78f12567c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.572089 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_028c96a4-38cf-4710-bcb2-a132836d77f1/openstack-network-exporter/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.579910 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_028c96a4-38cf-4710-bcb2-a132836d77f1/ovn-northd/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.902803 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5321fb79-5d79-4094-b314-8914ed41cfd5/ovsdbserver-nb/0.log" Oct 10 10:32:30 crc kubenswrapper[4669]: I1010 10:32:30.921207 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5321fb79-5d79-4094-b314-8914ed41cfd5/openstack-network-exporter/0.log" Oct 10 10:32:31 crc kubenswrapper[4669]: I1010 10:32:31.158151 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9f30537-9153-4862-97aa-31e4f5e57ecf/openstack-network-exporter/0.log" Oct 10 10:32:31 crc kubenswrapper[4669]: I1010 10:32:31.263678 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b9f30537-9153-4862-97aa-31e4f5e57ecf/ovsdbserver-sb/0.log" Oct 10 10:32:31 crc kubenswrapper[4669]: I1010 10:32:31.496888 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f557bc54-vbfbn_d9be36cc-2136-4285-8489-93149d0136c3/placement-api/0.log" Oct 10 10:32:31 crc kubenswrapper[4669]: I1010 10:32:31.622832 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f557bc54-vbfbn_d9be36cc-2136-4285-8489-93149d0136c3/placement-log/0.log" Oct 10 10:32:31 crc kubenswrapper[4669]: I1010 10:32:31.634081 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05416508-674b-44d8-9074-62776915ec92/setup-container/0.log" Oct 10 10:32:32 crc kubenswrapper[4669]: I1010 10:32:32.086490 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05416508-674b-44d8-9074-62776915ec92/setup-container/0.log" Oct 10 10:32:32 crc kubenswrapper[4669]: I1010 10:32:32.222607 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05416508-674b-44d8-9074-62776915ec92/rabbitmq/0.log" Oct 10 10:32:32 crc kubenswrapper[4669]: I1010 10:32:32.255282 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1172c8c-d2a5-4f42-b97a-87265385580b/setup-container/0.log" Oct 10 10:32:32 crc kubenswrapper[4669]: I1010 10:32:32.589950 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1172c8c-d2a5-4f42-b97a-87265385580b/setup-container/0.log" Oct 10 10:32:32 crc kubenswrapper[4669]: I1010 10:32:32.598769 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c1172c8c-d2a5-4f42-b97a-87265385580b/rabbitmq/0.log" Oct 10 10:32:32 crc kubenswrapper[4669]: I1010 10:32:32.732146 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-xv64c_f9938853-891c-4766-82ef-21e7b718dcc2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:32 crc kubenswrapper[4669]: I1010 10:32:32.877915 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ndzdh_5c0a91b3-706e-47df-8f2b-84733ad00593/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:33 crc kubenswrapper[4669]: I1010 10:32:33.023254 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-768lk_483acfce-519d-4164-a1fe-e8786ade906f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:33 crc kubenswrapper[4669]: I1010 10:32:33.328690 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dpnrk_5dc32d38-3a9b-48bb-93c4-76f51d215213/ssh-known-hosts-edpm-deployment/0.log" Oct 10 10:32:33 crc kubenswrapper[4669]: I1010 10:32:33.492925 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_732227c2-1b04-4af1-9c97-98e7dd2c4760/tempest-tests-tempest-tests-runner/0.log" Oct 10 10:32:33 crc kubenswrapper[4669]: I1010 10:32:33.526091 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_acf0f799-f812-4c43-8fc1-f018e701eebe/test-operator-logs-container/0.log" Oct 10 10:32:33 crc kubenswrapper[4669]: I1010 10:32:33.775097 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9pvnl_4feb106b-763c-422d-a1df-a95b354a7770/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 10 10:32:49 crc kubenswrapper[4669]: I1010 10:32:49.680925 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9c376674-4cf1-40c0-9e3c-c25e3fefad30/memcached/0.log" Oct 10 10:33:08 crc kubenswrapper[4669]: I1010 10:33:08.671645 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-sbc5v_affd00ea-eac5-4c5f-b452-c8bb0e17aba8/kube-rbac-proxy/0.log" Oct 10 10:33:08 crc kubenswrapper[4669]: I1010 10:33:08.727857 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-sbc5v_affd00ea-eac5-4c5f-b452-c8bb0e17aba8/manager/0.log" Oct 10 10:33:08 crc kubenswrapper[4669]: I1010 10:33:08.957156 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-jzps8_7dcc64bf-ac2d-4357-adcb-b6c146462464/kube-rbac-proxy/0.log" Oct 10 10:33:08 crc kubenswrapper[4669]: I1010 10:33:08.999903 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-jzps8_7dcc64bf-ac2d-4357-adcb-b6c146462464/manager/0.log" Oct 10 10:33:09 crc kubenswrapper[4669]: I1010 10:33:09.870382 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-pn8xd_1a5e08a1-2d18-4382-8129-be1376c40bde/kube-rbac-proxy/0.log" Oct 10 10:33:09 crc kubenswrapper[4669]: I1010 10:33:09.873699 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-pn8xd_1a5e08a1-2d18-4382-8129-be1376c40bde/manager/0.log" Oct 10 10:33:09 crc kubenswrapper[4669]: I1010 10:33:09.900939 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/util/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.074201 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/util/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.093672 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/pull/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.124711 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/pull/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.430964 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/pull/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.473546 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/extract/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.499162 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f60551a1fb7e79e9788cd8502fdb7f98c74e281ccbf8bf8c740740bbabqhdbn_236d5734-1ffa-4791-abf6-2fe59ade1eeb/util/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.699448 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-pfp4n_55f7e993-3dd8-4acf-9716-47cef7f73ef8/kube-rbac-proxy/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.854579 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-pdnnc_bd3dd74a-3c58-4785-854e-400c7ba726de/kube-rbac-proxy/0.log" Oct 10 10:33:10 crc kubenswrapper[4669]: I1010 10:33:10.860317 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-pfp4n_55f7e993-3dd8-4acf-9716-47cef7f73ef8/manager/0.log" Oct 10 10:33:11 crc kubenswrapper[4669]: I1010 10:33:11.015498 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-pdnnc_bd3dd74a-3c58-4785-854e-400c7ba726de/manager/0.log" Oct 10 10:33:11 crc kubenswrapper[4669]: I1010 10:33:11.850266 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-96278_909f202d-6571-4743-9314-26cc27e94c77/kube-rbac-proxy/0.log" Oct 10 10:33:11 crc kubenswrapper[4669]: I1010 10:33:11.851899 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-czldp_270b1e97-6178-44d9-8d9f-ed5e4dc02000/kube-rbac-proxy/0.log" Oct 10 10:33:11 crc kubenswrapper[4669]: I1010 10:33:11.859082 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-96278_909f202d-6571-4743-9314-26cc27e94c77/manager/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.109758 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-k5nxp_2132b635-c43e-4c4d-9074-37e81b6345f2/kube-rbac-proxy/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.232310 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-k5nxp_2132b635-c43e-4c4d-9074-37e81b6345f2/manager/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.264362 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-czldp_270b1e97-6178-44d9-8d9f-ed5e4dc02000/manager/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.417114 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-6zprx_27b6cee3-b0a3-4116-b9bb-299a81bba403/kube-rbac-proxy/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.554108 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-6zprx_27b6cee3-b0a3-4116-b9bb-299a81bba403/manager/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.766105 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-q52hz_6941897d-fd2e-4ff3-ad50-aeba82eb86c2/kube-rbac-proxy/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.771359 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-q52hz_6941897d-fd2e-4ff3-ad50-aeba82eb86c2/manager/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.810162 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-n72xm_82353170-7f06-43e4-966b-20d0d234f0f2/kube-rbac-proxy/0.log" Oct 10 10:33:12 crc kubenswrapper[4669]: I1010 10:33:12.908567 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-n72xm_82353170-7f06-43e4-966b-20d0d234f0f2/manager/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.054539 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-kwn4g_bdebd762-e639-4e50-820d-cd32885a1341/kube-rbac-proxy/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.136180 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-kwn4g_bdebd762-e639-4e50-820d-cd32885a1341/manager/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.221495 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-pdf8v_48aca3a8-684a-448f-ab17-b9d604d47484/kube-rbac-proxy/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.313701 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-wtsqj_43bf605f-c6c8-447b-9dc3-03be42fe7f10/kube-rbac-proxy/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.319437 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-pdf8v_48aca3a8-684a-448f-ab17-b9d604d47484/manager/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.417868 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-wtsqj_43bf605f-c6c8-447b-9dc3-03be42fe7f10/manager/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.578578 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv_73105b8d-0527-4510-886b-7203163993fd/kube-rbac-proxy/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.601520 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84c868ff4cn7tvv_73105b8d-0527-4510-886b-7203163993fd/manager/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.739200 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5698bb9464-flrhj_5f2708ab-b3ba-4556-9e11-7018b186975e/kube-rbac-proxy/0.log" Oct 10 10:33:13 crc kubenswrapper[4669]: I1010 10:33:13.881014 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-599bffcb5d-2bj6v_d6df673a-6169-4bcc-bf13-98eb9e3a20f9/kube-rbac-proxy/0.log" Oct 10 10:33:14 crc kubenswrapper[4669]: I1010 10:33:14.098038 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-599bffcb5d-2bj6v_d6df673a-6169-4bcc-bf13-98eb9e3a20f9/operator/0.log" Oct 10 10:33:14 crc kubenswrapper[4669]: I1010 10:33:14.280046 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2tjnm_0e034df9-3c3c-4b15-8d82-2a4179090ed7/registry-server/0.log" Oct 10 10:33:14 crc kubenswrapper[4669]: I1010 10:33:14.428719 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-r7ctx_29661927-099c-4549-8584-9390881170d5/kube-rbac-proxy/0.log" Oct 10 10:33:14 crc kubenswrapper[4669]: I1010 10:33:14.613952 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-r7ctx_29661927-099c-4549-8584-9390881170d5/manager/0.log" Oct 10 10:33:14 crc kubenswrapper[4669]: I1010 10:33:14.725124 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-4b4jb_316d31d2-d9a9-4fd7-a079-4620ad8c5e09/kube-rbac-proxy/0.log" Oct 10 10:33:14 crc kubenswrapper[4669]: I1010 10:33:14.974266 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-4b4jb_316d31d2-d9a9-4fd7-a079-4620ad8c5e09/manager/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.045868 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-gxpzs_637efcba-83b3-4f3a-93f0-1ce6257987e1/operator/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.259667 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-hjkht_db1620d5-baef-433a-b9e7-07a55004a68f/kube-rbac-proxy/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.291837 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5698bb9464-flrhj_5f2708ab-b3ba-4556-9e11-7018b186975e/manager/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.431306 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-hjkht_db1620d5-baef-433a-b9e7-07a55004a68f/manager/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.510504 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-j7fm7_f30a8280-3312-4333-8966-810b8dd3c889/kube-rbac-proxy/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.593426 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-j7fm7_f30a8280-3312-4333-8966-810b8dd3c889/manager/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.650011 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-b2p7l_684c8731-c863-49d2-a029-d9d69ae01f8d/kube-rbac-proxy/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.810364 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-b2p7l_684c8731-c863-49d2-a029-d9d69ae01f8d/manager/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.953359 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-cjtjr_5d2f190a-4eae-44e6-a5cb-e15abf2e037d/kube-rbac-proxy/0.log" Oct 10 10:33:15 crc kubenswrapper[4669]: I1010 10:33:15.973492 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-cjtjr_5d2f190a-4eae-44e6-a5cb-e15abf2e037d/manager/0.log" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.586779 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cjst2"] Oct 10 10:33:23 crc kubenswrapper[4669]: E1010 10:33:23.588489 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362ce889-52b7-42ba-8ee5-306532d627c1" containerName="container-00" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.588513 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="362ce889-52b7-42ba-8ee5-306532d627c1" containerName="container-00" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.588739 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="362ce889-52b7-42ba-8ee5-306532d627c1" containerName="container-00" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.590202 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.603919 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjst2"] Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.674705 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkxrw\" (UniqueName: \"kubernetes.io/projected/21e5712a-6f1b-4619-a3b3-3af647883ef8-kube-api-access-dkxrw\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.675183 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-catalog-content\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.675205 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-utilities\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.776783 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkxrw\" (UniqueName: \"kubernetes.io/projected/21e5712a-6f1b-4619-a3b3-3af647883ef8-kube-api-access-dkxrw\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.777132 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-catalog-content\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.777314 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-utilities\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.777748 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-catalog-content\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.777793 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-utilities\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.811695 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkxrw\" (UniqueName: \"kubernetes.io/projected/21e5712a-6f1b-4619-a3b3-3af647883ef8-kube-api-access-dkxrw\") pod \"certified-operators-cjst2\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:23 crc kubenswrapper[4669]: I1010 10:33:23.933003 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:24 crc kubenswrapper[4669]: I1010 10:33:24.276342 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:33:24 crc kubenswrapper[4669]: I1010 10:33:24.276856 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:33:24 crc kubenswrapper[4669]: I1010 10:33:24.476838 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjst2"] Oct 10 10:33:25 crc kubenswrapper[4669]: I1010 10:33:25.067041 4669 generic.go:334] "Generic (PLEG): container finished" podID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerID="16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991" exitCode=0 Oct 10 10:33:25 crc kubenswrapper[4669]: I1010 10:33:25.067222 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjst2" event={"ID":"21e5712a-6f1b-4619-a3b3-3af647883ef8","Type":"ContainerDied","Data":"16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991"} Oct 10 10:33:25 crc kubenswrapper[4669]: I1010 10:33:25.067451 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjst2" event={"ID":"21e5712a-6f1b-4619-a3b3-3af647883ef8","Type":"ContainerStarted","Data":"3e2cc847ab606513da752b93c83ebbf1eba06117a290663750478024652ee3ae"} Oct 10 10:33:26 crc kubenswrapper[4669]: I1010 10:33:26.108503 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjst2" event={"ID":"21e5712a-6f1b-4619-a3b3-3af647883ef8","Type":"ContainerStarted","Data":"a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3"} Oct 10 10:33:28 crc kubenswrapper[4669]: I1010 10:33:28.128458 4669 generic.go:334] "Generic (PLEG): container finished" podID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerID="a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3" exitCode=0 Oct 10 10:33:28 crc kubenswrapper[4669]: I1010 10:33:28.128550 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjst2" event={"ID":"21e5712a-6f1b-4619-a3b3-3af647883ef8","Type":"ContainerDied","Data":"a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3"} Oct 10 10:33:29 crc kubenswrapper[4669]: I1010 10:33:29.143091 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjst2" event={"ID":"21e5712a-6f1b-4619-a3b3-3af647883ef8","Type":"ContainerStarted","Data":"c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525"} Oct 10 10:33:29 crc kubenswrapper[4669]: I1010 10:33:29.171055 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cjst2" podStartSLOduration=2.508307473 podStartE2EDuration="6.171033392s" podCreationTimestamp="2025-10-10 10:33:23 +0000 UTC" firstStartedPulling="2025-10-10 10:33:25.069336138 +0000 UTC m=+4948.085354880" lastFinishedPulling="2025-10-10 10:33:28.732062047 +0000 UTC m=+4951.748080799" observedRunningTime="2025-10-10 10:33:29.163898313 +0000 UTC m=+4952.179917055" watchObservedRunningTime="2025-10-10 10:33:29.171033392 +0000 UTC m=+4952.187052134" Oct 10 10:33:33 crc kubenswrapper[4669]: I1010 10:33:33.933370 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:33 crc kubenswrapper[4669]: I1010 10:33:33.934150 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:34 crc kubenswrapper[4669]: I1010 10:33:34.592715 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:34 crc kubenswrapper[4669]: I1010 10:33:34.766729 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:34 crc kubenswrapper[4669]: I1010 10:33:34.853287 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjst2"] Oct 10 10:33:36 crc kubenswrapper[4669]: I1010 10:33:36.213709 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cjst2" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="registry-server" containerID="cri-o://c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525" gracePeriod=2 Oct 10 10:33:36 crc kubenswrapper[4669]: I1010 10:33:36.857182 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:36 crc kubenswrapper[4669]: I1010 10:33:36.988206 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-utilities\") pod \"21e5712a-6f1b-4619-a3b3-3af647883ef8\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " Oct 10 10:33:36 crc kubenswrapper[4669]: I1010 10:33:36.988648 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkxrw\" (UniqueName: \"kubernetes.io/projected/21e5712a-6f1b-4619-a3b3-3af647883ef8-kube-api-access-dkxrw\") pod \"21e5712a-6f1b-4619-a3b3-3af647883ef8\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " Oct 10 10:33:36 crc kubenswrapper[4669]: I1010 10:33:36.988786 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-catalog-content\") pod \"21e5712a-6f1b-4619-a3b3-3af647883ef8\" (UID: \"21e5712a-6f1b-4619-a3b3-3af647883ef8\") " Oct 10 10:33:36 crc kubenswrapper[4669]: I1010 10:33:36.989829 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-utilities" (OuterVolumeSpecName: "utilities") pod "21e5712a-6f1b-4619-a3b3-3af647883ef8" (UID: "21e5712a-6f1b-4619-a3b3-3af647883ef8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:33:36 crc kubenswrapper[4669]: I1010 10:33:36.996209 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e5712a-6f1b-4619-a3b3-3af647883ef8-kube-api-access-dkxrw" (OuterVolumeSpecName: "kube-api-access-dkxrw") pod "21e5712a-6f1b-4619-a3b3-3af647883ef8" (UID: "21e5712a-6f1b-4619-a3b3-3af647883ef8"). InnerVolumeSpecName "kube-api-access-dkxrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.039949 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21e5712a-6f1b-4619-a3b3-3af647883ef8" (UID: "21e5712a-6f1b-4619-a3b3-3af647883ef8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.091864 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkxrw\" (UniqueName: \"kubernetes.io/projected/21e5712a-6f1b-4619-a3b3-3af647883ef8-kube-api-access-dkxrw\") on node \"crc\" DevicePath \"\"" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.091908 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.091918 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21e5712a-6f1b-4619-a3b3-3af647883ef8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.230434 4669 generic.go:334] "Generic (PLEG): container finished" podID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerID="c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525" exitCode=0 Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.230487 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjst2" event={"ID":"21e5712a-6f1b-4619-a3b3-3af647883ef8","Type":"ContainerDied","Data":"c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525"} Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.230520 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjst2" event={"ID":"21e5712a-6f1b-4619-a3b3-3af647883ef8","Type":"ContainerDied","Data":"3e2cc847ab606513da752b93c83ebbf1eba06117a290663750478024652ee3ae"} Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.230539 4669 scope.go:117] "RemoveContainer" containerID="c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.230662 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjst2" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.271484 4669 scope.go:117] "RemoveContainer" containerID="a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.290149 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjst2"] Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.308438 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cjst2"] Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.310979 4669 scope.go:117] "RemoveContainer" containerID="16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.362131 4669 scope.go:117] "RemoveContainer" containerID="c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525" Oct 10 10:33:37 crc kubenswrapper[4669]: E1010 10:33:37.362762 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525\": container with ID starting with c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525 not found: ID does not exist" containerID="c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.362798 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525"} err="failed to get container status \"c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525\": rpc error: code = NotFound desc = could not find container \"c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525\": container with ID starting with c10238a5dffccdcc478e42578234f16c55b67533e09f94d16decff1b6c7bb525 not found: ID does not exist" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.362827 4669 scope.go:117] "RemoveContainer" containerID="a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3" Oct 10 10:33:37 crc kubenswrapper[4669]: E1010 10:33:37.363430 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3\": container with ID starting with a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3 not found: ID does not exist" containerID="a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.363463 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3"} err="failed to get container status \"a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3\": rpc error: code = NotFound desc = could not find container \"a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3\": container with ID starting with a561b4c109ab64a01c9247e744abb32b1e4dc3c67f284bafbafe187a110895a3 not found: ID does not exist" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.363481 4669 scope.go:117] "RemoveContainer" containerID="16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991" Oct 10 10:33:37 crc kubenswrapper[4669]: E1010 10:33:37.363797 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991\": container with ID starting with 16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991 not found: ID does not exist" containerID="16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.363828 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991"} err="failed to get container status \"16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991\": rpc error: code = NotFound desc = could not find container \"16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991\": container with ID starting with 16792d5bbff6c566b0f0d20fbfa02b20aa91b1a1c4b55c7facba466dba68f991 not found: ID does not exist" Oct 10 10:33:37 crc kubenswrapper[4669]: I1010 10:33:37.813255 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" path="/var/lib/kubelet/pods/21e5712a-6f1b-4619-a3b3-3af647883ef8/volumes" Oct 10 10:33:42 crc kubenswrapper[4669]: I1010 10:33:42.548778 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-f6dtf_f273bc2e-26d7-40fe-9150-fb582fb14618/control-plane-machine-set-operator/0.log" Oct 10 10:33:42 crc kubenswrapper[4669]: I1010 10:33:42.830503 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9lpfb_e57d53b2-5487-4820-aa61-9070eb866eae/kube-rbac-proxy/0.log" Oct 10 10:33:42 crc kubenswrapper[4669]: I1010 10:33:42.852244 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9lpfb_e57d53b2-5487-4820-aa61-9070eb866eae/machine-api-operator/0.log" Oct 10 10:33:54 crc kubenswrapper[4669]: I1010 10:33:54.275185 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:33:54 crc kubenswrapper[4669]: I1010 10:33:54.276096 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:33:58 crc kubenswrapper[4669]: I1010 10:33:58.236051 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-45l5f_a9cc3175-a61f-44c4-8668-6270126b84b7/cert-manager-cainjector/0.log" Oct 10 10:33:58 crc kubenswrapper[4669]: I1010 10:33:58.240931 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-tqwjt_4ab4da17-72b9-45f6-94c3-75905972615a/cert-manager-controller/0.log" Oct 10 10:33:58 crc kubenswrapper[4669]: I1010 10:33:58.485865 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4258m_302f9809-912d-4e36-b0d5-7bdbae70071a/cert-manager-webhook/0.log" Oct 10 10:34:12 crc kubenswrapper[4669]: I1010 10:34:12.016751 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-brznz_5e2fc45c-0a2e-4c3c-82ce-98ba90a5bce0/nmstate-console-plugin/0.log" Oct 10 10:34:12 crc kubenswrapper[4669]: I1010 10:34:12.268333 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-jlckz_35a376a3-b98d-4c9d-b7a8-633759d368ea/kube-rbac-proxy/0.log" Oct 10 10:34:12 crc kubenswrapper[4669]: I1010 10:34:12.271343 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-v5rbb_27433356-4791-4652-b258-aaf3d6deebef/nmstate-handler/0.log" Oct 10 10:34:12 crc kubenswrapper[4669]: I1010 10:34:12.402301 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-jlckz_35a376a3-b98d-4c9d-b7a8-633759d368ea/nmstate-metrics/0.log" Oct 10 10:34:12 crc kubenswrapper[4669]: I1010 10:34:12.523329 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-qwxrv_1b2674b7-3907-4f5a-b3b8-cc6945e4f004/nmstate-operator/0.log" Oct 10 10:34:12 crc kubenswrapper[4669]: I1010 10:34:12.649957 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-dm9zp_9ec015c7-dd0b-46e0-a5b9-456bf97c6a49/nmstate-webhook/0.log" Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.274685 4669 patch_prober.go:28] interesting pod/machine-config-daemon-x6v7p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.275408 4669 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.275471 4669 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.276677 4669 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d"} pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.276769 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" containerName="machine-config-daemon" containerID="cri-o://4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" gracePeriod=600 Oct 10 10:34:24 crc kubenswrapper[4669]: E1010 10:34:24.415164 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.702770 4669 generic.go:334] "Generic (PLEG): container finished" podID="addb758f-1f34-4793-af67-1a54167543b9" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" exitCode=0 Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.702878 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerDied","Data":"4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d"} Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.703282 4669 scope.go:117] "RemoveContainer" containerID="bc11853ac15a382d5d96f25f2a58173703d5b2ed2201893ca7b2e71e54ea802f" Oct 10 10:34:24 crc kubenswrapper[4669]: I1010 10:34:24.704150 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:34:24 crc kubenswrapper[4669]: E1010 10:34:24.704568 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:34:33 crc kubenswrapper[4669]: I1010 10:34:33.402830 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xh4cl_0eeead67-0b6d-48c7-8756-50e1c1eb5f0b/kube-rbac-proxy/0.log" Oct 10 10:34:33 crc kubenswrapper[4669]: I1010 10:34:33.521840 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-xh4cl_0eeead67-0b6d-48c7-8756-50e1c1eb5f0b/controller/0.log" Oct 10 10:34:33 crc kubenswrapper[4669]: I1010 10:34:33.665111 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:34:33 crc kubenswrapper[4669]: I1010 10:34:33.931940 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:34:33 crc kubenswrapper[4669]: I1010 10:34:33.996810 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.007375 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.056044 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.202382 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.294920 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.295353 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.399715 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.653251 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-frr-files/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.712579 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-metrics/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.719656 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/controller/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.751419 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/cp-reloader/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.962908 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/kube-rbac-proxy/0.log" Oct 10 10:34:34 crc kubenswrapper[4669]: I1010 10:34:34.992063 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/frr-metrics/0.log" Oct 10 10:34:35 crc kubenswrapper[4669]: I1010 10:34:35.094601 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/kube-rbac-proxy-frr/0.log" Oct 10 10:34:35 crc kubenswrapper[4669]: I1010 10:34:35.302192 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/reloader/0.log" Oct 10 10:34:35 crc kubenswrapper[4669]: I1010 10:34:35.502216 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-rgcll_8bd73190-2f9a-45df-8e2d-8579238bf4a3/frr-k8s-webhook-server/0.log" Oct 10 10:34:35 crc kubenswrapper[4669]: I1010 10:34:35.796951 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:34:35 crc kubenswrapper[4669]: E1010 10:34:35.797161 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:34:35 crc kubenswrapper[4669]: I1010 10:34:35.877289 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79d96cb9f5-vkfb4_5cda559e-d9fe-4651-b77d-d8c7d38e6a21/manager/0.log" Oct 10 10:34:36 crc kubenswrapper[4669]: I1010 10:34:36.099485 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7dd78c579-pxx42_06827887-c1ea-4839-9ed3-a4c8813d81ac/webhook-server/0.log" Oct 10 10:34:36 crc kubenswrapper[4669]: I1010 10:34:36.262977 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-78hzw_ec1f1180-0070-43bd-9ee3-095e8e2379f0/kube-rbac-proxy/0.log" Oct 10 10:34:36 crc kubenswrapper[4669]: I1010 10:34:36.661133 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-rp4mn_3f8d268c-022e-42b5-8e08-dbdefd895e23/frr/0.log" Oct 10 10:34:36 crc kubenswrapper[4669]: I1010 10:34:36.945923 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-78hzw_ec1f1180-0070-43bd-9ee3-095e8e2379f0/speaker/0.log" Oct 10 10:34:50 crc kubenswrapper[4669]: I1010 10:34:50.795634 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:34:50 crc kubenswrapper[4669]: E1010 10:34:50.796901 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:34:51 crc kubenswrapper[4669]: I1010 10:34:51.397700 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/util/0.log" Oct 10 10:34:51 crc kubenswrapper[4669]: I1010 10:34:51.656163 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/util/0.log" Oct 10 10:34:51 crc kubenswrapper[4669]: I1010 10:34:51.674308 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/pull/0.log" Oct 10 10:34:51 crc kubenswrapper[4669]: I1010 10:34:51.696222 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/pull/0.log" Oct 10 10:34:51 crc kubenswrapper[4669]: I1010 10:34:51.887192 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/util/0.log" Oct 10 10:34:51 crc kubenswrapper[4669]: I1010 10:34:51.889909 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/extract/0.log" Oct 10 10:34:51 crc kubenswrapper[4669]: I1010 10:34:51.923158 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2wm658_5e1484dd-9b20-445b-9aaa-30884d9fd717/pull/0.log" Oct 10 10:34:52 crc kubenswrapper[4669]: I1010 10:34:52.200526 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-utilities/0.log" Oct 10 10:34:52 crc kubenswrapper[4669]: I1010 10:34:52.494339 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-utilities/0.log" Oct 10 10:34:52 crc kubenswrapper[4669]: I1010 10:34:52.568453 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-content/0.log" Oct 10 10:34:52 crc kubenswrapper[4669]: I1010 10:34:52.649001 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-content/0.log" Oct 10 10:34:52 crc kubenswrapper[4669]: I1010 10:34:52.764177 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-utilities/0.log" Oct 10 10:34:52 crc kubenswrapper[4669]: I1010 10:34:52.888067 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/extract-content/0.log" Oct 10 10:34:53 crc kubenswrapper[4669]: I1010 10:34:53.184086 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-utilities/0.log" Oct 10 10:34:53 crc kubenswrapper[4669]: I1010 10:34:53.473913 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5b7lm_e8ff4b09-5116-45de-9f97-f87832000566/registry-server/0.log" Oct 10 10:34:53 crc kubenswrapper[4669]: I1010 10:34:53.578556 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-utilities/0.log" Oct 10 10:34:53 crc kubenswrapper[4669]: I1010 10:34:53.579724 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-content/0.log" Oct 10 10:34:53 crc kubenswrapper[4669]: I1010 10:34:53.627212 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-content/0.log" Oct 10 10:34:53 crc kubenswrapper[4669]: I1010 10:34:53.872339 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-content/0.log" Oct 10 10:34:53 crc kubenswrapper[4669]: I1010 10:34:53.883096 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/extract-utilities/0.log" Oct 10 10:34:54 crc kubenswrapper[4669]: I1010 10:34:54.298041 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/util/0.log" Oct 10 10:34:54 crc kubenswrapper[4669]: I1010 10:34:54.586734 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/util/0.log" Oct 10 10:34:54 crc kubenswrapper[4669]: I1010 10:34:54.686834 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/pull/0.log" Oct 10 10:34:54 crc kubenswrapper[4669]: I1010 10:34:54.755307 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/pull/0.log" Oct 10 10:34:54 crc kubenswrapper[4669]: I1010 10:34:54.804686 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fq9ss_9a078913-33a1-429a-850e-400f184f3cc8/registry-server/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.019306 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/pull/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.054258 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/util/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.117798 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cz2plp_15e5bfc4-8089-4e7f-9c96-66e5090b8dcc/extract/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.321336 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wf49x_90231823-44be-4740-bef0-c98a048f9b05/marketplace-operator/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.462091 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-utilities/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.566463 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-utilities/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.617309 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-content/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.622056 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-content/0.log" Oct 10 10:34:55 crc kubenswrapper[4669]: I1010 10:34:55.928281 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-utilities/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.000837 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/extract-content/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.156275 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-tltpl_0aa2e009-01b6-46c5-bc27-41c9c62e46c5/registry-server/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.240358 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-utilities/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.450145 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-content/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.481349 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-utilities/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.533600 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-content/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.774295 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-utilities/0.log" Oct 10 10:34:56 crc kubenswrapper[4669]: I1010 10:34:56.782159 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/extract-content/0.log" Oct 10 10:34:57 crc kubenswrapper[4669]: I1010 10:34:57.263391 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-phbdn_1ce843ca-00a5-4c8a-abbb-dd9fb719f331/registry-server/0.log" Oct 10 10:35:04 crc kubenswrapper[4669]: I1010 10:35:04.795963 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:35:04 crc kubenswrapper[4669]: E1010 10:35:04.796709 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:35:15 crc kubenswrapper[4669]: I1010 10:35:15.795381 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:35:15 crc kubenswrapper[4669]: E1010 10:35:15.796409 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:35:28 crc kubenswrapper[4669]: I1010 10:35:28.801133 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:35:28 crc kubenswrapper[4669]: E1010 10:35:28.802561 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:35:35 crc kubenswrapper[4669]: E1010 10:35:35.855308 4669 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.53:59920->38.102.83.53:45881: write tcp 38.102.83.53:59920->38.102.83.53:45881: write: broken pipe Oct 10 10:35:41 crc kubenswrapper[4669]: I1010 10:35:41.800980 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:35:41 crc kubenswrapper[4669]: E1010 10:35:41.803099 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:35:54 crc kubenswrapper[4669]: I1010 10:35:54.796106 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:35:54 crc kubenswrapper[4669]: E1010 10:35:54.797040 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:36:06 crc kubenswrapper[4669]: I1010 10:36:06.796718 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:36:06 crc kubenswrapper[4669]: E1010 10:36:06.798210 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:36:20 crc kubenswrapper[4669]: I1010 10:36:20.795841 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:36:20 crc kubenswrapper[4669]: E1010 10:36:20.796816 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:36:34 crc kubenswrapper[4669]: I1010 10:36:34.796767 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:36:34 crc kubenswrapper[4669]: E1010 10:36:34.797920 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.079530 4669 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jclsb"] Oct 10 10:36:36 crc kubenswrapper[4669]: E1010 10:36:36.081015 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="extract-utilities" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.081046 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="extract-utilities" Oct 10 10:36:36 crc kubenswrapper[4669]: E1010 10:36:36.081061 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="registry-server" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.081069 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="registry-server" Oct 10 10:36:36 crc kubenswrapper[4669]: E1010 10:36:36.081133 4669 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="extract-content" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.081144 4669 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="extract-content" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.081385 4669 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e5712a-6f1b-4619-a3b3-3af647883ef8" containerName="registry-server" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.083631 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.114335 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jclsb"] Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.185430 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-utilities\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.185511 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-catalog-content\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.185725 4669 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvgj5\" (UniqueName: \"kubernetes.io/projected/55013059-c6d8-450a-b8fe-aee65a8e90e8-kube-api-access-vvgj5\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.288386 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvgj5\" (UniqueName: \"kubernetes.io/projected/55013059-c6d8-450a-b8fe-aee65a8e90e8-kube-api-access-vvgj5\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.288437 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-utilities\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.288502 4669 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-catalog-content\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.289170 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-catalog-content\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.289200 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-utilities\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.319282 4669 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvgj5\" (UniqueName: \"kubernetes.io/projected/55013059-c6d8-450a-b8fe-aee65a8e90e8-kube-api-access-vvgj5\") pod \"redhat-operators-jclsb\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.415825 4669 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:36 crc kubenswrapper[4669]: I1010 10:36:36.982109 4669 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jclsb"] Oct 10 10:36:37 crc kubenswrapper[4669]: I1010 10:36:37.157826 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jclsb" event={"ID":"55013059-c6d8-450a-b8fe-aee65a8e90e8","Type":"ContainerStarted","Data":"b8232d16db2b426ce4b1a21a0f5243796394fe273e7d94923755da2080073e48"} Oct 10 10:36:38 crc kubenswrapper[4669]: I1010 10:36:38.171859 4669 generic.go:334] "Generic (PLEG): container finished" podID="55013059-c6d8-450a-b8fe-aee65a8e90e8" containerID="4ac35756dd58b308047b47559ebb5bd6049fe4e62209a6971a3426a1497eb13f" exitCode=0 Oct 10 10:36:38 crc kubenswrapper[4669]: I1010 10:36:38.172109 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jclsb" event={"ID":"55013059-c6d8-450a-b8fe-aee65a8e90e8","Type":"ContainerDied","Data":"4ac35756dd58b308047b47559ebb5bd6049fe4e62209a6971a3426a1497eb13f"} Oct 10 10:36:38 crc kubenswrapper[4669]: I1010 10:36:38.175818 4669 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 10:36:40 crc kubenswrapper[4669]: I1010 10:36:40.194854 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jclsb" event={"ID":"55013059-c6d8-450a-b8fe-aee65a8e90e8","Type":"ContainerStarted","Data":"f04a5017e5a7a544077a994211d177ddf5876e2f77b68366cb6311dda3bbaf5f"} Oct 10 10:36:43 crc kubenswrapper[4669]: I1010 10:36:43.229460 4669 generic.go:334] "Generic (PLEG): container finished" podID="55013059-c6d8-450a-b8fe-aee65a8e90e8" containerID="f04a5017e5a7a544077a994211d177ddf5876e2f77b68366cb6311dda3bbaf5f" exitCode=0 Oct 10 10:36:43 crc kubenswrapper[4669]: I1010 10:36:43.229578 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jclsb" event={"ID":"55013059-c6d8-450a-b8fe-aee65a8e90e8","Type":"ContainerDied","Data":"f04a5017e5a7a544077a994211d177ddf5876e2f77b68366cb6311dda3bbaf5f"} Oct 10 10:36:45 crc kubenswrapper[4669]: I1010 10:36:45.261000 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jclsb" event={"ID":"55013059-c6d8-450a-b8fe-aee65a8e90e8","Type":"ContainerStarted","Data":"934e28389075b9f5950ebf223e13f5410f7c390e9578aadf354f8b8025b1e120"} Oct 10 10:36:45 crc kubenswrapper[4669]: I1010 10:36:45.289103 4669 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jclsb" podStartSLOduration=3.645548216 podStartE2EDuration="9.289082779s" podCreationTimestamp="2025-10-10 10:36:36 +0000 UTC" firstStartedPulling="2025-10-10 10:36:38.175478474 +0000 UTC m=+5141.191497216" lastFinishedPulling="2025-10-10 10:36:43.819013037 +0000 UTC m=+5146.835031779" observedRunningTime="2025-10-10 10:36:45.283982356 +0000 UTC m=+5148.300001108" watchObservedRunningTime="2025-10-10 10:36:45.289082779 +0000 UTC m=+5148.305101521" Oct 10 10:36:46 crc kubenswrapper[4669]: I1010 10:36:46.416885 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:46 crc kubenswrapper[4669]: I1010 10:36:46.417410 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:36:47 crc kubenswrapper[4669]: I1010 10:36:47.469012 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jclsb" podUID="55013059-c6d8-450a-b8fe-aee65a8e90e8" containerName="registry-server" probeResult="failure" output=< Oct 10 10:36:47 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 10:36:47 crc kubenswrapper[4669]: > Oct 10 10:36:47 crc kubenswrapper[4669]: I1010 10:36:47.814842 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:36:47 crc kubenswrapper[4669]: E1010 10:36:47.815135 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:36:57 crc kubenswrapper[4669]: I1010 10:36:57.473439 4669 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jclsb" podUID="55013059-c6d8-450a-b8fe-aee65a8e90e8" containerName="registry-server" probeResult="failure" output=< Oct 10 10:36:57 crc kubenswrapper[4669]: timeout: failed to connect service ":50051" within 1s Oct 10 10:36:57 crc kubenswrapper[4669]: > Oct 10 10:37:00 crc kubenswrapper[4669]: I1010 10:37:00.796166 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:37:00 crc kubenswrapper[4669]: E1010 10:37:00.797231 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:37:06 crc kubenswrapper[4669]: I1010 10:37:06.480030 4669 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:37:06 crc kubenswrapper[4669]: I1010 10:37:06.561753 4669 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:37:07 crc kubenswrapper[4669]: I1010 10:37:07.282325 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jclsb"] Oct 10 10:37:07 crc kubenswrapper[4669]: I1010 10:37:07.504199 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jclsb" podUID="55013059-c6d8-450a-b8fe-aee65a8e90e8" containerName="registry-server" containerID="cri-o://934e28389075b9f5950ebf223e13f5410f7c390e9578aadf354f8b8025b1e120" gracePeriod=2 Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.521191 4669 generic.go:334] "Generic (PLEG): container finished" podID="55013059-c6d8-450a-b8fe-aee65a8e90e8" containerID="934e28389075b9f5950ebf223e13f5410f7c390e9578aadf354f8b8025b1e120" exitCode=0 Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.521418 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jclsb" event={"ID":"55013059-c6d8-450a-b8fe-aee65a8e90e8","Type":"ContainerDied","Data":"934e28389075b9f5950ebf223e13f5410f7c390e9578aadf354f8b8025b1e120"} Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.521629 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jclsb" event={"ID":"55013059-c6d8-450a-b8fe-aee65a8e90e8","Type":"ContainerDied","Data":"b8232d16db2b426ce4b1a21a0f5243796394fe273e7d94923755da2080073e48"} Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.521658 4669 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8232d16db2b426ce4b1a21a0f5243796394fe273e7d94923755da2080073e48" Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.824394 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.837246 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-utilities\") pod \"55013059-c6d8-450a-b8fe-aee65a8e90e8\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.837360 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-catalog-content\") pod \"55013059-c6d8-450a-b8fe-aee65a8e90e8\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.837422 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvgj5\" (UniqueName: \"kubernetes.io/projected/55013059-c6d8-450a-b8fe-aee65a8e90e8-kube-api-access-vvgj5\") pod \"55013059-c6d8-450a-b8fe-aee65a8e90e8\" (UID: \"55013059-c6d8-450a-b8fe-aee65a8e90e8\") " Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.838223 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-utilities" (OuterVolumeSpecName: "utilities") pod "55013059-c6d8-450a-b8fe-aee65a8e90e8" (UID: "55013059-c6d8-450a-b8fe-aee65a8e90e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.853959 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55013059-c6d8-450a-b8fe-aee65a8e90e8-kube-api-access-vvgj5" (OuterVolumeSpecName: "kube-api-access-vvgj5") pod "55013059-c6d8-450a-b8fe-aee65a8e90e8" (UID: "55013059-c6d8-450a-b8fe-aee65a8e90e8"). InnerVolumeSpecName "kube-api-access-vvgj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.940500 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvgj5\" (UniqueName: \"kubernetes.io/projected/55013059-c6d8-450a-b8fe-aee65a8e90e8-kube-api-access-vvgj5\") on node \"crc\" DevicePath \"\"" Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.940963 4669 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 10:37:08 crc kubenswrapper[4669]: I1010 10:37:08.946015 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55013059-c6d8-450a-b8fe-aee65a8e90e8" (UID: "55013059-c6d8-450a-b8fe-aee65a8e90e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:37:09 crc kubenswrapper[4669]: I1010 10:37:09.042730 4669 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55013059-c6d8-450a-b8fe-aee65a8e90e8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 10:37:09 crc kubenswrapper[4669]: I1010 10:37:09.531334 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jclsb" Oct 10 10:37:09 crc kubenswrapper[4669]: I1010 10:37:09.581549 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jclsb"] Oct 10 10:37:09 crc kubenswrapper[4669]: I1010 10:37:09.591521 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jclsb"] Oct 10 10:37:09 crc kubenswrapper[4669]: I1010 10:37:09.812713 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55013059-c6d8-450a-b8fe-aee65a8e90e8" path="/var/lib/kubelet/pods/55013059-c6d8-450a-b8fe-aee65a8e90e8/volumes" Oct 10 10:37:11 crc kubenswrapper[4669]: I1010 10:37:11.796387 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:37:11 crc kubenswrapper[4669]: E1010 10:37:11.799091 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:37:24 crc kubenswrapper[4669]: I1010 10:37:24.796385 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:37:24 crc kubenswrapper[4669]: E1010 10:37:24.798574 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:37:25 crc kubenswrapper[4669]: I1010 10:37:25.565093 4669 scope.go:117] "RemoveContainer" containerID="7d726e68c009b93af97c9effbba36df188fe8b75fc3b1835d1a0af37491b0246" Oct 10 10:37:35 crc kubenswrapper[4669]: I1010 10:37:35.795725 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:37:35 crc kubenswrapper[4669]: E1010 10:37:35.796677 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:37:39 crc kubenswrapper[4669]: I1010 10:37:39.855878 4669 generic.go:334] "Generic (PLEG): container finished" podID="d49fc2f9-06e0-409c-9829-65375fad7e5b" containerID="e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0" exitCode=0 Oct 10 10:37:39 crc kubenswrapper[4669]: I1010 10:37:39.856105 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" event={"ID":"d49fc2f9-06e0-409c-9829-65375fad7e5b","Type":"ContainerDied","Data":"e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0"} Oct 10 10:37:39 crc kubenswrapper[4669]: I1010 10:37:39.859714 4669 scope.go:117] "RemoveContainer" containerID="e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0" Oct 10 10:37:40 crc kubenswrapper[4669]: I1010 10:37:40.010258 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xkr7t_must-gather-pzqxv_d49fc2f9-06e0-409c-9829-65375fad7e5b/gather/0.log" Oct 10 10:37:47 crc kubenswrapper[4669]: I1010 10:37:47.813725 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:37:47 crc kubenswrapper[4669]: E1010 10:37:47.814995 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.004505 4669 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xkr7t/must-gather-pzqxv"] Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.005456 4669 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" podUID="d49fc2f9-06e0-409c-9829-65375fad7e5b" containerName="copy" containerID="cri-o://0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34" gracePeriod=2 Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.019562 4669 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xkr7t/must-gather-pzqxv"] Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.457240 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xkr7t_must-gather-pzqxv_d49fc2f9-06e0-409c-9829-65375fad7e5b/copy/0.log" Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.458117 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.551202 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d49fc2f9-06e0-409c-9829-65375fad7e5b-must-gather-output\") pod \"d49fc2f9-06e0-409c-9829-65375fad7e5b\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.551551 4669 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj9wx\" (UniqueName: \"kubernetes.io/projected/d49fc2f9-06e0-409c-9829-65375fad7e5b-kube-api-access-zj9wx\") pod \"d49fc2f9-06e0-409c-9829-65375fad7e5b\" (UID: \"d49fc2f9-06e0-409c-9829-65375fad7e5b\") " Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.565914 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49fc2f9-06e0-409c-9829-65375fad7e5b-kube-api-access-zj9wx" (OuterVolumeSpecName: "kube-api-access-zj9wx") pod "d49fc2f9-06e0-409c-9829-65375fad7e5b" (UID: "d49fc2f9-06e0-409c-9829-65375fad7e5b"). InnerVolumeSpecName "kube-api-access-zj9wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.654660 4669 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj9wx\" (UniqueName: \"kubernetes.io/projected/d49fc2f9-06e0-409c-9829-65375fad7e5b-kube-api-access-zj9wx\") on node \"crc\" DevicePath \"\"" Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.772187 4669 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d49fc2f9-06e0-409c-9829-65375fad7e5b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d49fc2f9-06e0-409c-9829-65375fad7e5b" (UID: "d49fc2f9-06e0-409c-9829-65375fad7e5b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.815328 4669 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49fc2f9-06e0-409c-9829-65375fad7e5b" path="/var/lib/kubelet/pods/d49fc2f9-06e0-409c-9829-65375fad7e5b/volumes" Oct 10 10:37:53 crc kubenswrapper[4669]: I1010 10:37:53.864227 4669 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d49fc2f9-06e0-409c-9829-65375fad7e5b-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.042461 4669 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xkr7t_must-gather-pzqxv_d49fc2f9-06e0-409c-9829-65375fad7e5b/copy/0.log" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.043301 4669 generic.go:334] "Generic (PLEG): container finished" podID="d49fc2f9-06e0-409c-9829-65375fad7e5b" containerID="0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34" exitCode=143 Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.043382 4669 scope.go:117] "RemoveContainer" containerID="0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.043416 4669 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xkr7t/must-gather-pzqxv" Oct 10 10:37:54 crc kubenswrapper[4669]: E1010 10:37:54.069358 4669 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd49fc2f9_06e0_409c_9829_65375fad7e5b.slice/crio-9058563d688c5ff3a36d3bfae29495a097724cd2698d8752535c4ae891a1fac4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd49fc2f9_06e0_409c_9829_65375fad7e5b.slice\": RecentStats: unable to find data in memory cache]" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.074033 4669 scope.go:117] "RemoveContainer" containerID="e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.156466 4669 scope.go:117] "RemoveContainer" containerID="0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34" Oct 10 10:37:54 crc kubenswrapper[4669]: E1010 10:37:54.158156 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34\": container with ID starting with 0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34 not found: ID does not exist" containerID="0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.158224 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34"} err="failed to get container status \"0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34\": rpc error: code = NotFound desc = could not find container \"0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34\": container with ID starting with 0a9b341a4199e3117538ee2c5877c9413a9a3d9dee92d596adf1cb2d6394ba34 not found: ID does not exist" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.158273 4669 scope.go:117] "RemoveContainer" containerID="e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0" Oct 10 10:37:54 crc kubenswrapper[4669]: E1010 10:37:54.159562 4669 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0\": container with ID starting with e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0 not found: ID does not exist" containerID="e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0" Oct 10 10:37:54 crc kubenswrapper[4669]: I1010 10:37:54.159661 4669 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0"} err="failed to get container status \"e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0\": rpc error: code = NotFound desc = could not find container \"e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0\": container with ID starting with e4c0443055192071a60bce7595005f110ed40668586e9d281df60c52cb5045e0 not found: ID does not exist" Oct 10 10:37:58 crc kubenswrapper[4669]: I1010 10:37:58.796634 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:37:58 crc kubenswrapper[4669]: E1010 10:37:58.797643 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:38:10 crc kubenswrapper[4669]: I1010 10:38:10.795643 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:38:10 crc kubenswrapper[4669]: E1010 10:38:10.796861 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:38:26 crc kubenswrapper[4669]: I1010 10:38:26.795910 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:38:26 crc kubenswrapper[4669]: E1010 10:38:26.797043 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:38:38 crc kubenswrapper[4669]: I1010 10:38:38.796904 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:38:38 crc kubenswrapper[4669]: E1010 10:38:38.798184 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:38:50 crc kubenswrapper[4669]: I1010 10:38:50.796735 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:38:50 crc kubenswrapper[4669]: E1010 10:38:50.797817 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:39:05 crc kubenswrapper[4669]: I1010 10:39:05.796089 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:39:05 crc kubenswrapper[4669]: E1010 10:39:05.797230 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:39:16 crc kubenswrapper[4669]: I1010 10:39:16.795860 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:39:16 crc kubenswrapper[4669]: E1010 10:39:16.797139 4669 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-x6v7p_openshift-machine-config-operator(addb758f-1f34-4793-af67-1a54167543b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" podUID="addb758f-1f34-4793-af67-1a54167543b9" Oct 10 10:39:27 crc kubenswrapper[4669]: I1010 10:39:27.812911 4669 scope.go:117] "RemoveContainer" containerID="4e761df40096c9218dd7e0d9bda1a5b55af7de6922c165d890fb5b77d8202d2d" Oct 10 10:39:28 crc kubenswrapper[4669]: I1010 10:39:28.048066 4669 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-x6v7p" event={"ID":"addb758f-1f34-4793-af67-1a54167543b9","Type":"ContainerStarted","Data":"9ca6fcdc307b98f2b400520bf4f8911a2938be3cf853d80832860a39069090e6"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515072161222024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015072161223017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015072146150016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015072146151015457 5ustar corecore